[ 466.717751] env[62460]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=62460) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 466.718172] env[62460]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=62460) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 466.718300] env[62460]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=62460) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 466.718574] env[62460]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 466.815207] env[62460]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=62460) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 466.825659] env[62460]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=62460) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 467.430240] env[62460]: INFO nova.virt.driver [None req-61512f4d-99ea-46db-afee-6b2135913912 None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 467.500449] env[62460]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 467.500688] env[62460]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 467.500785] env[62460]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=62460) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 470.539083] env[62460]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-ff8506a7-4991-4efa-ab88-de335012d99e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 470.554806] env[62460]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=62460) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 470.555032] env[62460]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-60a0709e-c677-43d5-913a-981ef5564bef {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 470.593971] env[62460]: INFO oslo_vmware.api [-] Successfully established new session; session ID is 6d19c. [ 470.594592] env[62460]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.094s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 470.594794] env[62460]: INFO nova.virt.vmwareapi.driver [None req-61512f4d-99ea-46db-afee-6b2135913912 None None] VMware vCenter version: 7.0.3 [ 470.598152] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54098179-cf4b-4ee0-b9a0-f33d0a102729 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 470.619804] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33409336-e027-41b2-bd2e-8e2efc61ea15 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 470.625517] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4452ab8f-39af-4d0b-bea0-53baab89e5d6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 470.631769] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ca75a80-4743-413a-ab07-59d670aa0da2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 470.644429] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-334a339f-458a-4ed2-b055-2b44e026b1bc {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 470.650161] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b31a7564-0f85-41cd-879e-a1729115b8cc {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 470.679736] env[62460]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-1ac7dcd1-c274-41ad-bd48-61e0c0f683bb {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 470.684695] env[62460]: DEBUG nova.virt.vmwareapi.driver [None req-61512f4d-99ea-46db-afee-6b2135913912 None None] Extension org.openstack.compute already exists. {{(pid=62460) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:226}} [ 470.687298] env[62460]: INFO nova.compute.provider_config [None req-61512f4d-99ea-46db-afee-6b2135913912 None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 471.190559] env[62460]: DEBUG nova.context [None req-61512f4d-99ea-46db-afee-6b2135913912 None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),412ba090-791d-4f9b-8c74-84937e35f9ac(cell1) {{(pid=62460) load_cells /opt/stack/nova/nova/context.py:464}} [ 471.192617] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 471.192847] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 471.193859] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 471.193963] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] Acquiring lock "412ba090-791d-4f9b-8c74-84937e35f9ac" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 471.194168] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] Lock "412ba090-791d-4f9b-8c74-84937e35f9ac" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 471.195185] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] Lock "412ba090-791d-4f9b-8c74-84937e35f9ac" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 471.215569] env[62460]: INFO dbcounter [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] Registered counter for database nova_cell0 [ 471.223641] env[62460]: INFO dbcounter [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] Registered counter for database nova_cell1 [ 471.227150] env[62460]: DEBUG oslo_db.sqlalchemy.engines [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=62460) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 471.227767] env[62460]: DEBUG oslo_db.sqlalchemy.engines [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=62460) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 471.232360] env[62460]: ERROR nova.db.main.api [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 471.232360] env[62460]: result = function(*args, **kwargs) [ 471.232360] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 471.232360] env[62460]: return func(*args, **kwargs) [ 471.232360] env[62460]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 471.232360] env[62460]: result = fn(*args, **kwargs) [ 471.232360] env[62460]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 471.232360] env[62460]: return f(*args, **kwargs) [ 471.232360] env[62460]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 471.232360] env[62460]: return db.service_get_minimum_version(context, binaries) [ 471.232360] env[62460]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 471.232360] env[62460]: _check_db_access() [ 471.232360] env[62460]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 471.232360] env[62460]: stacktrace = ''.join(traceback.format_stack()) [ 471.232360] env[62460]: [ 471.233357] env[62460]: ERROR nova.db.main.api [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 471.233357] env[62460]: result = function(*args, **kwargs) [ 471.233357] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 471.233357] env[62460]: return func(*args, **kwargs) [ 471.233357] env[62460]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 471.233357] env[62460]: result = fn(*args, **kwargs) [ 471.233357] env[62460]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 471.233357] env[62460]: return f(*args, **kwargs) [ 471.233357] env[62460]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 471.233357] env[62460]: return db.service_get_minimum_version(context, binaries) [ 471.233357] env[62460]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 471.233357] env[62460]: _check_db_access() [ 471.233357] env[62460]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 471.233357] env[62460]: stacktrace = ''.join(traceback.format_stack()) [ 471.233357] env[62460]: [ 471.233763] env[62460]: WARNING nova.objects.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 471.233888] env[62460]: WARNING nova.objects.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] Failed to get minimum service version for cell 412ba090-791d-4f9b-8c74-84937e35f9ac [ 471.234330] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] Acquiring lock "singleton_lock" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 471.234496] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] Acquired lock "singleton_lock" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 471.234741] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] Releasing lock "singleton_lock" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 471.235084] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] Full set of CONF: {{(pid=62460) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 471.235234] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] ******************************************************************************** {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 471.235362] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] Configuration options gathered from: {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 471.235513] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 471.235713] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 471.235843] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] ================================================================================ {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 471.236074] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] allow_resize_to_same_host = True {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.236255] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] arq_binding_timeout = 300 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.236387] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] backdoor_port = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.236515] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] backdoor_socket = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.236680] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] block_device_allocate_retries = 60 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.236840] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] block_device_allocate_retries_interval = 3 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.237013] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cert = self.pem {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.237193] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.237369] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] compute_monitors = [] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.237532] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] config_dir = [] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.237700] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] config_drive_format = iso9660 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.237834] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.237998] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] config_source = [] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.238215] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] console_host = devstack {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.238389] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] control_exchange = nova {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.238550] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cpu_allocation_ratio = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.238711] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] daemon = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.238879] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] debug = True {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.239051] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] default_access_ip_network_name = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.239227] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] default_availability_zone = nova {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.239384] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] default_ephemeral_format = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.239546] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] default_green_pool_size = 1000 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.239785] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.239954] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] default_schedule_zone = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.240170] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] disk_allocation_ratio = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.240322] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] enable_new_services = True {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.240502] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] enabled_apis = ['osapi_compute'] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.240670] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] enabled_ssl_apis = [] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.240833] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] flat_injected = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.240994] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] force_config_drive = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.241176] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] force_raw_images = True {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.241349] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] graceful_shutdown_timeout = 5 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.241514] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] heal_instance_info_cache_interval = 60 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.241791] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] host = cpu-1 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.241910] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] initial_cpu_allocation_ratio = 4.0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.242089] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] initial_disk_allocation_ratio = 1.0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.242262] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] initial_ram_allocation_ratio = 1.0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.242477] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.242645] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] instance_build_timeout = 0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.242805] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] instance_delete_interval = 300 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.242973] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] instance_format = [instance: %(uuid)s] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.243157] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] instance_name_template = instance-%08x {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.243323] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] instance_usage_audit = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.243494] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] instance_usage_audit_period = month {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.243660] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.243825] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] instances_path = /opt/stack/data/nova/instances {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.243991] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] internal_service_availability_zone = internal {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.244165] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] key = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.244328] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] live_migration_retry_count = 30 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.244497] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] log_color = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.244665] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] log_config_append = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.244832] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.244992] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] log_dir = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.245165] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] log_file = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.245297] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] log_options = True {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.245457] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] log_rotate_interval = 1 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.245627] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] log_rotate_interval_type = days {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.245793] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] log_rotation_type = none {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.245923] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.246067] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.246243] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.246410] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.246539] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.246701] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] long_rpc_timeout = 1800 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.246859] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] max_concurrent_builds = 10 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.247027] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] max_concurrent_live_migrations = 1 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.247194] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] max_concurrent_snapshots = 5 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.247355] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] max_local_block_devices = 3 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.247514] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] max_logfile_count = 30 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.247671] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] max_logfile_size_mb = 200 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.247830] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] maximum_instance_delete_attempts = 5 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.247994] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] metadata_listen = 0.0.0.0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.248201] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] metadata_listen_port = 8775 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.248380] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] metadata_workers = 2 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.248544] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] migrate_max_retries = -1 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.248712] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] mkisofs_cmd = genisoimage {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.248921] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] my_block_storage_ip = 10.180.1.21 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.249066] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] my_ip = 10.180.1.21 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.249237] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] network_allocate_retries = 0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.249415] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.249581] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] osapi_compute_listen = 0.0.0.0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.249743] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] osapi_compute_listen_port = 8774 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.249909] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] osapi_compute_unique_server_name_scope = {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.250113] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] osapi_compute_workers = 2 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.250300] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] password_length = 12 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.250465] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] periodic_enable = True {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.250624] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] periodic_fuzzy_delay = 60 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.250791] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] pointer_model = usbtablet {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.250957] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] preallocate_images = none {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.251151] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] publish_errors = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.251291] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] pybasedir = /opt/stack/nova {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.251452] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] ram_allocation_ratio = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.251611] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] rate_limit_burst = 0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.251776] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] rate_limit_except_level = CRITICAL {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.251935] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] rate_limit_interval = 0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.252108] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] reboot_timeout = 0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.252271] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] reclaim_instance_interval = 0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.252427] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] record = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.252594] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] reimage_timeout_per_gb = 60 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.252756] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] report_interval = 120 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.252916] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] rescue_timeout = 0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.253086] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] reserved_host_cpus = 0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.253249] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] reserved_host_disk_mb = 0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.253406] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] reserved_host_memory_mb = 512 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.253561] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] reserved_huge_pages = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.253719] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] resize_confirm_window = 0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.253874] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] resize_fs_using_block_device = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.254040] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] resume_guests_state_on_host_boot = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.254213] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.254373] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] rpc_response_timeout = 60 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.254531] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] run_external_periodic_tasks = True {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.254696] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] running_deleted_instance_action = reap {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.254854] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] running_deleted_instance_poll_interval = 1800 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.255015] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] running_deleted_instance_timeout = 0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.255179] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] scheduler_instance_sync_interval = 120 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.255346] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] service_down_time = 720 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.255512] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] servicegroup_driver = db {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.255665] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] shell_completion = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.255821] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] shelved_offload_time = 0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.255976] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] shelved_poll_interval = 3600 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.256154] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] shutdown_timeout = 0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.256315] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] source_is_ipv6 = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.256471] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] ssl_only = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.256710] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.256878] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] sync_power_state_interval = 600 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.257052] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] sync_power_state_pool_size = 1000 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.257224] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] syslog_log_facility = LOG_USER {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.257379] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] tempdir = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.257535] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] timeout_nbd = 10 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.257701] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] transport_url = **** {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.257858] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] update_resources_interval = 0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.258025] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] use_cow_images = True {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.258222] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] use_eventlog = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.258388] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] use_journal = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.258601] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] use_json = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.258863] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] use_rootwrap_daemon = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.259139] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] use_stderr = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.259352] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] use_syslog = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.259523] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vcpu_pin_set = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.259696] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vif_plugging_is_fatal = True {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.259866] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vif_plugging_timeout = 300 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.260048] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] virt_mkfs = [] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.260239] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] volume_usage_poll_interval = 0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.260408] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] watch_log_file = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.260577] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] web = /usr/share/spice-html5 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.260764] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] os_brick.lock_path = /opt/stack/data/n-cpu-1 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.260943] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.261130] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] os_brick.wait_mpath_device_interval = 1 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.261306] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_concurrency.disable_process_locking = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.261863] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.262074] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.262258] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.262437] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_messaging_metrics.metrics_process_name = {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.262614] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.262784] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.262969] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] api.auth_strategy = keystone {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.263158] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] api.compute_link_prefix = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.263339] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.263516] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] api.dhcp_domain = novalocal {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.263688] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] api.enable_instance_password = True {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.263857] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] api.glance_link_prefix = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.264038] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.264224] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] api.instance_list_cells_batch_strategy = distributed {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.264390] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] api.instance_list_per_project_cells = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.264554] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] api.list_records_by_skipping_down_cells = True {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.264717] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] api.local_metadata_per_cell = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.264888] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] api.max_limit = 1000 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.265069] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] api.metadata_cache_expiration = 15 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.265251] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] api.neutron_default_tenant_id = default {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.265423] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] api.response_validation = warn {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.265596] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] api.use_neutron_default_nets = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.265767] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.265932] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] api.vendordata_dynamic_failure_fatal = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.266130] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.266293] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] api.vendordata_dynamic_ssl_certfile = {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.266467] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] api.vendordata_dynamic_targets = [] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.266633] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] api.vendordata_jsonfile_path = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.266814] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] api.vendordata_providers = ['StaticJSON'] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.267017] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cache.backend = dogpile.cache.memcached {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.267192] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cache.backend_argument = **** {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.267368] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cache.config_prefix = cache.oslo {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.267537] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cache.dead_timeout = 60.0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.267704] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cache.debug_cache_backend = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.267868] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cache.enable_retry_client = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.268046] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cache.enable_socket_keepalive = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.268252] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cache.enabled = True {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.268426] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cache.enforce_fips_mode = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.268594] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cache.expiration_time = 600 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.268759] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cache.hashclient_retry_attempts = 2 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.268927] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cache.hashclient_retry_delay = 1.0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.269108] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cache.memcache_dead_retry = 300 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.269276] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cache.memcache_password = **** {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.269444] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.269605] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.269768] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cache.memcache_pool_maxsize = 10 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.269930] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cache.memcache_pool_unused_timeout = 60 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.270132] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cache.memcache_sasl_enabled = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.270330] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cache.memcache_servers = ['localhost:11211'] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.270504] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cache.memcache_socket_timeout = 1.0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.270669] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cache.memcache_username = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.270838] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cache.proxies = [] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.271015] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cache.redis_db = 0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.271238] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cache.redis_password = **** {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.271425] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cache.redis_sentinel_service_name = mymaster {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.271609] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.271782] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cache.redis_server = localhost:6379 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.271951] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cache.redis_socket_timeout = 1.0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.272132] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cache.redis_username = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.272313] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cache.retry_attempts = 2 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.272476] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cache.retry_delay = 0.0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.272640] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cache.socket_keepalive_count = 1 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.272802] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cache.socket_keepalive_idle = 1 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.272966] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cache.socket_keepalive_interval = 1 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.273144] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cache.tls_allowed_ciphers = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.273314] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cache.tls_cafile = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.273473] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cache.tls_certfile = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.273638] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cache.tls_enabled = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.273799] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cache.tls_keyfile = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.273971] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cinder.auth_section = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.274162] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cinder.auth_type = password {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.274333] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cinder.cafile = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.274547] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cinder.catalog_info = volumev3::publicURL {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.274738] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cinder.certfile = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.274911] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cinder.collect_timing = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.275092] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cinder.cross_az_attach = True {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.275269] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cinder.debug = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.275435] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cinder.endpoint_template = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.275602] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cinder.http_retries = 3 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.275768] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cinder.insecure = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.275931] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cinder.keyfile = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.276119] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cinder.os_region_name = RegionOne {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.276295] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cinder.split_loggers = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.276460] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cinder.timeout = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.276637] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.276801] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] compute.cpu_dedicated_set = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.276961] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] compute.cpu_shared_set = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.277145] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] compute.image_type_exclude_list = [] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.277324] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] compute.live_migration_wait_for_vif_plug = True {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.277486] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] compute.max_concurrent_disk_ops = 0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.277650] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] compute.max_disk_devices_to_attach = -1 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.277816] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.277990] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.278204] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] compute.resource_provider_association_refresh = 300 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.278384] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.278553] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] compute.shutdown_retry_interval = 10 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.278738] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.278922] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] conductor.workers = 2 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.279121] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] console.allowed_origins = [] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.279290] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] console.ssl_ciphers = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.279464] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] console.ssl_minimum_version = default {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.279637] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] consoleauth.enforce_session_timeout = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.279811] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] consoleauth.token_ttl = 600 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.279981] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cyborg.cafile = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.280462] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cyborg.certfile = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.280462] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cyborg.collect_timing = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.280522] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cyborg.connect_retries = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.280652] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cyborg.connect_retry_delay = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.280817] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cyborg.endpoint_override = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.280986] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cyborg.insecure = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.281163] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cyborg.keyfile = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.281327] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cyborg.max_version = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.281486] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cyborg.min_version = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.281645] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cyborg.region_name = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.281804] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cyborg.retriable_status_codes = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.281965] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cyborg.service_name = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.282151] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cyborg.service_type = accelerator {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.282318] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cyborg.split_loggers = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.282477] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cyborg.status_code_retries = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.282637] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cyborg.status_code_retry_delay = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.282794] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cyborg.timeout = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.282975] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.283154] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] cyborg.version = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.283340] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] database.backend = sqlalchemy {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.283513] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] database.connection = **** {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.283683] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] database.connection_debug = 0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.283855] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] database.connection_parameters = {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.284036] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] database.connection_recycle_time = 3600 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.284210] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] database.connection_trace = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.284377] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] database.db_inc_retry_interval = True {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.285468] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] database.db_max_retries = 20 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.285468] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] database.db_max_retry_interval = 10 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.285468] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] database.db_retry_interval = 1 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.285468] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] database.max_overflow = 50 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.285468] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] database.max_pool_size = 5 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.285468] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] database.max_retries = 10 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.285672] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] database.mysql_sql_mode = TRADITIONAL {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.285672] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] database.mysql_wsrep_sync_wait = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.285818] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] database.pool_timeout = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.285977] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] database.retry_interval = 10 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.286160] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] database.slave_connection = **** {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.286326] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] database.sqlite_synchronous = True {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.287180] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] database.use_db_reconnect = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.287180] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] api_database.backend = sqlalchemy {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.287180] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] api_database.connection = **** {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.287180] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] api_database.connection_debug = 0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.287180] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] api_database.connection_parameters = {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.287359] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] api_database.connection_recycle_time = 3600 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.287480] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] api_database.connection_trace = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.287644] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] api_database.db_inc_retry_interval = True {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.287808] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] api_database.db_max_retries = 20 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.287974] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] api_database.db_max_retry_interval = 10 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.288176] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] api_database.db_retry_interval = 1 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.288354] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] api_database.max_overflow = 50 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.292013] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] api_database.max_pool_size = 5 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.292013] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] api_database.max_retries = 10 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.292013] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.292013] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] api_database.mysql_wsrep_sync_wait = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.292013] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] api_database.pool_timeout = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.292013] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] api_database.retry_interval = 10 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.292013] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] api_database.slave_connection = **** {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.292243] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] api_database.sqlite_synchronous = True {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.292243] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] devices.enabled_mdev_types = [] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.292243] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.292243] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] ephemeral_storage_encryption.default_format = luks {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.292243] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] ephemeral_storage_encryption.enabled = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.292243] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] ephemeral_storage_encryption.key_size = 512 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.292243] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] glance.api_servers = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.292419] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] glance.cafile = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.292419] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] glance.certfile = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.292419] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] glance.collect_timing = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.292419] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] glance.connect_retries = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.292419] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] glance.connect_retry_delay = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.292419] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] glance.debug = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.292419] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] glance.default_trusted_certificate_ids = [] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.292590] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] glance.enable_certificate_validation = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.292590] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] glance.enable_rbd_download = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.292590] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] glance.endpoint_override = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.292590] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] glance.insecure = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.292693] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] glance.keyfile = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.292868] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] glance.max_version = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.292928] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] glance.min_version = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.293105] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] glance.num_retries = 3 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.293272] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] glance.rbd_ceph_conf = {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.293431] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] glance.rbd_connect_timeout = 5 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.293597] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] glance.rbd_pool = {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.293761] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] glance.rbd_user = {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.293919] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] glance.region_name = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.294099] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] glance.retriable_status_codes = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.294261] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] glance.service_name = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.294427] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] glance.service_type = image {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.294585] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] glance.split_loggers = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.296315] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] glance.status_code_retries = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.296315] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] glance.status_code_retry_delay = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.296315] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] glance.timeout = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.296315] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.296315] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] glance.verify_glance_signatures = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.296315] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] glance.version = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.296315] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] guestfs.debug = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.296693] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] mks.enabled = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.296693] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.296693] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] image_cache.manager_interval = 2400 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.296693] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] image_cache.precache_concurrency = 1 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.296864] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] image_cache.remove_unused_base_images = True {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.296909] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.297090] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.297276] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] image_cache.subdirectory_name = _base {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.297455] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] ironic.api_max_retries = 60 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.297621] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] ironic.api_retry_interval = 2 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.297781] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] ironic.auth_section = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.297942] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] ironic.auth_type = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.298140] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] ironic.cafile = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.298320] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] ironic.certfile = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.298486] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] ironic.collect_timing = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.298650] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] ironic.conductor_group = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.298809] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] ironic.connect_retries = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.298967] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] ironic.connect_retry_delay = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.299145] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] ironic.endpoint_override = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.299311] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] ironic.insecure = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.299471] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] ironic.keyfile = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.299630] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] ironic.max_version = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.299785] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] ironic.min_version = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.299949] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] ironic.peer_list = [] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.300150] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] ironic.region_name = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.300331] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] ironic.retriable_status_codes = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.300544] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] ironic.serial_console_state_timeout = 10 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.300744] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] ironic.service_name = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.300923] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] ironic.service_type = baremetal {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.301107] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] ironic.shard = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.301276] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] ironic.split_loggers = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.301439] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] ironic.status_code_retries = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.301600] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] ironic.status_code_retry_delay = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.301760] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] ironic.timeout = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.301941] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.302120] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] ironic.version = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.302312] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.302526] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] key_manager.fixed_key = **** {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.302729] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.302896] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] barbican.barbican_api_version = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.303072] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] barbican.barbican_endpoint = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.303253] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] barbican.barbican_endpoint_type = public {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.303414] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] barbican.barbican_region_name = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.303576] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] barbican.cafile = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.303736] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] barbican.certfile = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.303900] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] barbican.collect_timing = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.304077] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] barbican.insecure = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.304245] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] barbican.keyfile = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.304410] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] barbican.number_of_retries = 60 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.304572] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] barbican.retry_delay = 1 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.304736] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] barbican.send_service_user_token = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.304898] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] barbican.split_loggers = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.305073] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] barbican.timeout = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.305244] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] barbican.verify_ssl = True {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.305406] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] barbican.verify_ssl_path = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.305575] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] barbican_service_user.auth_section = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.305741] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] barbican_service_user.auth_type = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.305902] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] barbican_service_user.cafile = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.306078] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] barbican_service_user.certfile = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.306249] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] barbican_service_user.collect_timing = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.306413] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] barbican_service_user.insecure = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.306573] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] barbican_service_user.keyfile = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.306737] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] barbican_service_user.split_loggers = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.306896] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] barbican_service_user.timeout = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.307076] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vault.approle_role_id = **** {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.307244] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vault.approle_secret_id = **** {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.307414] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vault.kv_mountpoint = secret {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.307576] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vault.kv_path = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.307743] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vault.kv_version = 2 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.307906] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vault.namespace = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.308100] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vault.root_token_id = **** {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.308288] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vault.ssl_ca_crt_file = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.308461] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vault.timeout = 60.0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.308627] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vault.use_ssl = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.308798] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.308972] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] keystone.auth_section = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.309166] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] keystone.auth_type = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.309329] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] keystone.cafile = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.309491] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] keystone.certfile = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.309656] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] keystone.collect_timing = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.309819] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] keystone.connect_retries = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.309981] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] keystone.connect_retry_delay = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.310185] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] keystone.endpoint_override = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.310361] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] keystone.insecure = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.310524] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] keystone.keyfile = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.310686] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] keystone.max_version = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.310845] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] keystone.min_version = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.311011] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] keystone.region_name = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.311182] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] keystone.retriable_status_codes = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.311345] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] keystone.service_name = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.311516] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] keystone.service_type = identity {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.311680] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] keystone.split_loggers = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.311841] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] keystone.status_code_retries = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.312007] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] keystone.status_code_retry_delay = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.312178] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] keystone.timeout = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.312364] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.312524] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] keystone.version = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.312726] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.connection_uri = {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.312889] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.cpu_mode = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.313070] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.cpu_model_extra_flags = [] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.313245] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.cpu_models = [] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.313416] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.cpu_power_governor_high = performance {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.313586] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.cpu_power_governor_low = powersave {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.313751] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.cpu_power_management = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.313923] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.314102] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.device_detach_attempts = 8 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.314273] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.device_detach_timeout = 20 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.314439] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.disk_cachemodes = [] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.314599] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.disk_prefix = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.314763] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.enabled_perf_events = [] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.314927] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.file_backed_memory = 0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.315106] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.gid_maps = [] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.315272] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.hw_disk_discard = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.315431] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.hw_machine_type = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.315603] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.images_rbd_ceph_conf = {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.315770] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.315934] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.316118] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.images_rbd_glance_store_name = {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.316294] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.images_rbd_pool = rbd {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.316465] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.images_type = default {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.316624] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.images_volume_group = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.316786] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.inject_key = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.316948] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.inject_partition = -2 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.317126] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.inject_password = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.317294] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.iscsi_iface = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.317455] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.iser_use_multipath = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.317620] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.live_migration_bandwidth = 0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.317784] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.live_migration_completion_timeout = 800 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.317946] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.live_migration_downtime = 500 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.318143] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.live_migration_downtime_delay = 75 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.318323] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.live_migration_downtime_steps = 10 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.318488] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.live_migration_inbound_addr = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.318652] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.live_migration_permit_auto_converge = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.318820] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.live_migration_permit_post_copy = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.318984] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.live_migration_scheme = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.319177] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.live_migration_timeout_action = abort {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.319347] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.live_migration_tunnelled = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.319507] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.live_migration_uri = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.319671] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.live_migration_with_native_tls = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.319833] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.max_queues = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.319996] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.mem_stats_period_seconds = 10 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.320278] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.320454] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.nfs_mount_options = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.320760] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.320936] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.num_aoe_discover_tries = 3 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.321127] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.num_iser_scan_tries = 5 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.321294] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.num_memory_encrypted_guests = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.321460] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.num_nvme_discover_tries = 5 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.321625] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.num_pcie_ports = 0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.321793] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.num_volume_scan_tries = 5 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.321960] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.pmem_namespaces = [] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.322136] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.quobyte_client_cfg = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.322454] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.322642] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.rbd_connect_timeout = 5 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.322813] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.322980] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.323160] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.rbd_secret_uuid = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.323324] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.rbd_user = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.323491] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.realtime_scheduler_priority = 1 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.323664] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.remote_filesystem_transport = ssh {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.323826] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.rescue_image_id = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.323986] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.rescue_kernel_id = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.324163] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.rescue_ramdisk_id = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.324336] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.rng_dev_path = /dev/urandom {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.324495] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.rx_queue_size = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.324661] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.smbfs_mount_options = {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.324935] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.325129] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.snapshot_compression = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.325304] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.snapshot_image_format = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.325526] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.325697] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.sparse_logical_volumes = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.325863] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.swtpm_enabled = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.326046] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.swtpm_group = tss {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.326228] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.swtpm_user = tss {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.326399] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.sysinfo_serial = unique {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.326560] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.tb_cache_size = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.326720] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.tx_queue_size = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.326886] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.uid_maps = [] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.327064] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.use_virtio_for_bridges = True {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.327243] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.virt_type = kvm {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.327414] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.volume_clear = zero {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.327581] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.volume_clear_size = 0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.327748] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.volume_use_multipath = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.327911] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.vzstorage_cache_path = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.328104] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.328287] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.vzstorage_mount_group = qemu {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.328456] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.vzstorage_mount_opts = [] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.328627] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.328908] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.329106] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.vzstorage_mount_user = stack {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.329283] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.329463] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] neutron.auth_section = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.329643] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] neutron.auth_type = password {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.329808] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] neutron.cafile = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.329971] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] neutron.certfile = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.330182] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] neutron.collect_timing = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.330358] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] neutron.connect_retries = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.330521] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] neutron.connect_retry_delay = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.330693] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] neutron.default_floating_pool = public {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.330855] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] neutron.endpoint_override = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.331031] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] neutron.extension_sync_interval = 600 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.331201] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] neutron.http_retries = 3 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.331365] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] neutron.insecure = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.331528] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] neutron.keyfile = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.331687] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] neutron.max_version = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.331857] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] neutron.metadata_proxy_shared_secret = **** {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.332028] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] neutron.min_version = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.332204] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] neutron.ovs_bridge = br-int {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.332383] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] neutron.physnets = [] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.332576] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] neutron.region_name = RegionOne {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.332743] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] neutron.retriable_status_codes = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.332917] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] neutron.service_metadata_proxy = True {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.333095] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] neutron.service_name = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.333272] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] neutron.service_type = network {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.333436] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] neutron.split_loggers = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.333595] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] neutron.status_code_retries = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.333752] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] neutron.status_code_retry_delay = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.333911] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] neutron.timeout = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.334105] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.334274] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] neutron.version = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.334447] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] notifications.bdms_in_notifications = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.334624] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] notifications.default_level = INFO {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.334797] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] notifications.notification_format = unversioned {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.334962] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] notifications.notify_on_state_change = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.335156] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.335334] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] pci.alias = [] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.335504] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] pci.device_spec = [] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.335669] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] pci.report_in_placement = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.335846] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] placement.auth_section = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.336035] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] placement.auth_type = password {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.336217] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] placement.auth_url = http://10.180.1.21/identity {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.336418] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] placement.cafile = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.336593] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] placement.certfile = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.336761] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] placement.collect_timing = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.336923] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] placement.connect_retries = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.337097] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] placement.connect_retry_delay = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.337266] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] placement.default_domain_id = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.337428] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] placement.default_domain_name = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.337590] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] placement.domain_id = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.337751] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] placement.domain_name = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.337912] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] placement.endpoint_override = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.338088] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] placement.insecure = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.338256] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] placement.keyfile = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.338415] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] placement.max_version = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.338573] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] placement.min_version = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.338741] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] placement.password = **** {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.338902] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] placement.project_domain_id = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.339083] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] placement.project_domain_name = Default {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.339262] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] placement.project_id = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.339441] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] placement.project_name = service {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.339614] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] placement.region_name = RegionOne {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.339790] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] placement.retriable_status_codes = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.339955] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] placement.service_name = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.340172] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] placement.service_type = placement {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.340354] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] placement.split_loggers = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.340519] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] placement.status_code_retries = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.340684] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] placement.status_code_retry_delay = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.340847] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] placement.system_scope = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.341028] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] placement.timeout = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.341186] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] placement.trust_id = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.341348] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] placement.user_domain_id = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.341522] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] placement.user_domain_name = Default {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.341682] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] placement.user_id = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.341858] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] placement.username = nova {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.342055] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.342228] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] placement.version = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.342425] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] quota.cores = 20 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.342602] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] quota.count_usage_from_placement = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.342786] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.342950] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] quota.injected_file_content_bytes = 10240 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.343137] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] quota.injected_file_path_length = 255 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.343311] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] quota.injected_files = 5 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.343479] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] quota.instances = 10 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.343645] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] quota.key_pairs = 100 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.343810] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] quota.metadata_items = 128 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.343978] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] quota.ram = 51200 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.344154] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] quota.recheck_quota = True {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.344331] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] quota.server_group_members = 10 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.344499] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] quota.server_groups = 10 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.344677] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.344843] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.345022] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] scheduler.image_metadata_prefilter = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.345186] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.345357] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] scheduler.max_attempts = 3 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.345522] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] scheduler.max_placement_results = 1000 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.345687] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.345852] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] scheduler.query_placement_for_image_type_support = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.346024] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.346207] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] scheduler.workers = 2 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.346387] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.346562] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.346744] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.346914] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.347097] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.347270] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.347437] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.347629] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.347799] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] filter_scheduler.host_subset_size = 1 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.347979] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.348183] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] filter_scheduler.image_properties_default_architecture = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.348355] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.348525] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] filter_scheduler.isolated_hosts = [] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.348692] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] filter_scheduler.isolated_images = [] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.348857] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] filter_scheduler.max_instances_per_host = 50 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.349032] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.349204] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.349372] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] filter_scheduler.pci_in_placement = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.349535] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.349698] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.349862] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.350034] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.350232] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.350405] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.350572] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] filter_scheduler.track_instance_changes = True {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.350752] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.350924] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] metrics.required = True {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.351103] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] metrics.weight_multiplier = 1.0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.351272] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] metrics.weight_of_unavailable = -10000.0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.351440] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] metrics.weight_setting = [] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.351755] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.351929] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] serial_console.enabled = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.352121] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] serial_console.port_range = 10000:20000 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.352298] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.352501] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.352682] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] serial_console.serialproxy_port = 6083 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.352854] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] service_user.auth_section = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.353038] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] service_user.auth_type = password {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.353209] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] service_user.cafile = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.353372] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] service_user.certfile = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.353534] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] service_user.collect_timing = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.353695] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] service_user.insecure = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.353852] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] service_user.keyfile = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.354034] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] service_user.send_service_user_token = True {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.354205] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] service_user.split_loggers = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.354381] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] service_user.timeout = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.354550] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] spice.agent_enabled = True {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.354713] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] spice.enabled = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.355035] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.355232] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] spice.html5proxy_host = 0.0.0.0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.355404] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] spice.html5proxy_port = 6082 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.355566] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] spice.image_compression = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.355723] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] spice.jpeg_compression = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.355881] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] spice.playback_compression = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.356053] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] spice.require_secure = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.356228] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] spice.server_listen = 127.0.0.1 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.356397] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.356555] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] spice.streaming_mode = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.356712] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] spice.zlib_compression = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.356876] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] upgrade_levels.baseapi = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.357058] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] upgrade_levels.compute = auto {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.357225] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] upgrade_levels.conductor = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.357382] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] upgrade_levels.scheduler = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.357547] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vendordata_dynamic_auth.auth_section = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.357709] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vendordata_dynamic_auth.auth_type = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.357868] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vendordata_dynamic_auth.cafile = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.358062] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vendordata_dynamic_auth.certfile = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.358245] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vendordata_dynamic_auth.collect_timing = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.358411] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vendordata_dynamic_auth.insecure = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.358571] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vendordata_dynamic_auth.keyfile = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.358733] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vendordata_dynamic_auth.split_loggers = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.358892] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vendordata_dynamic_auth.timeout = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.359080] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vmware.api_retry_count = 10 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.359253] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vmware.ca_file = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.359427] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vmware.cache_prefix = devstack-image-cache {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.359595] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vmware.cluster_name = testcl1 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.359761] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vmware.connection_pool_size = 10 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.359920] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vmware.console_delay_seconds = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.360150] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vmware.datastore_regex = ^datastore.* {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.360380] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.360562] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vmware.host_password = **** {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.360732] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vmware.host_port = 443 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.360901] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vmware.host_username = administrator@vsphere.local {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.361087] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vmware.insecure = True {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.361257] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vmware.integration_bridge = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.361422] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vmware.maximum_objects = 100 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.361584] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vmware.pbm_default_policy = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.361746] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vmware.pbm_enabled = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.361903] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vmware.pbm_wsdl_location = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.362087] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.362250] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vmware.serial_port_proxy_uri = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.362424] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vmware.serial_port_service_uri = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.362603] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vmware.task_poll_interval = 0.5 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.362777] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vmware.use_linked_clone = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.362946] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vmware.vnc_keymap = en-us {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.363129] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vmware.vnc_port = 5900 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.363299] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vmware.vnc_port_total = 10000 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.363486] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vnc.auth_schemes = ['none'] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.363661] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vnc.enabled = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.363948] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.364148] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.364323] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vnc.novncproxy_port = 6080 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.364499] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vnc.server_listen = 127.0.0.1 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.364668] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.364828] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vnc.vencrypt_ca_certs = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.364985] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vnc.vencrypt_client_cert = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.365158] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vnc.vencrypt_client_key = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.365338] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.365502] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] workarounds.disable_deep_image_inspection = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.365661] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] workarounds.disable_fallback_pcpu_query = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.365818] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] workarounds.disable_group_policy_check_upcall = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.365977] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.366155] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] workarounds.disable_rootwrap = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.366318] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] workarounds.enable_numa_live_migration = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.366478] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.366638] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.366797] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] workarounds.handle_virt_lifecycle_events = True {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.366960] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] workarounds.libvirt_disable_apic = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.367137] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] workarounds.never_download_image_if_on_rbd = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.367305] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.367495] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.367716] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.367895] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.368072] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.368244] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.368406] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.368568] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.368737] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.368925] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.369111] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] wsgi.client_socket_timeout = 900 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.369283] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] wsgi.default_pool_size = 1000 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.369448] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] wsgi.keep_alive = True {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.369614] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] wsgi.max_header_line = 16384 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.369776] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] wsgi.secure_proxy_ssl_header = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.369937] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] wsgi.ssl_ca_file = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.370136] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] wsgi.ssl_cert_file = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.370317] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] wsgi.ssl_key_file = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.370487] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] wsgi.tcp_keepidle = 600 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.370670] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.370843] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] zvm.ca_file = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.371017] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] zvm.cloud_connector_url = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.371315] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.371495] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] zvm.reachable_timeout = 300 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.371681] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_policy.enforce_new_defaults = True {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.372101] env[62460]: WARNING oslo_config.cfg [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 471.372316] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_policy.enforce_scope = True {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.372526] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_policy.policy_default_rule = default {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.372725] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.372908] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_policy.policy_file = policy.yaml {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.373106] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.373281] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.373454] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.373614] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.373780] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.373953] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.374153] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.374341] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] profiler.connection_string = messaging:// {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.374513] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] profiler.enabled = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.374688] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] profiler.es_doc_type = notification {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.374859] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] profiler.es_scroll_size = 10000 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.375045] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] profiler.es_scroll_time = 2m {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.375221] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] profiler.filter_error_trace = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.375400] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] profiler.hmac_keys = **** {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.375571] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] profiler.sentinel_service_name = mymaster {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.375741] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] profiler.socket_timeout = 0.1 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.375909] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] profiler.trace_requests = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.376087] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] profiler.trace_sqlalchemy = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.376279] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] profiler_jaeger.process_tags = {} {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.376447] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] profiler_jaeger.service_name_prefix = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.376613] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] profiler_otlp.service_name_prefix = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.376781] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] remote_debug.host = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.376943] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] remote_debug.port = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.377140] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.377312] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.377482] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.377666] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.377826] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.378015] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.378206] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.378379] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.378545] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.378722] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_messaging_rabbit.hostname = devstack {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.378883] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.379069] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.379249] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.379466] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.379641] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.380672] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.380672] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.380672] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.380672] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.380827] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.381064] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.381314] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.381577] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.381813] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.382168] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.382285] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.382471] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.382696] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.383023] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.383174] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_messaging_rabbit.ssl = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.383395] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.383656] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.383826] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.384074] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.384371] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_messaging_rabbit.ssl_version = {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.384516] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.384751] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.384961] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_messaging_notifications.retry = -1 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.385220] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.385461] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_messaging_notifications.transport_url = **** {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.385582] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_limit.auth_section = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.385811] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_limit.auth_type = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.386021] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_limit.cafile = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.386088] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_limit.certfile = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.386252] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_limit.collect_timing = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.386434] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_limit.connect_retries = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.386690] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_limit.connect_retry_delay = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.386767] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_limit.endpoint_id = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.386940] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_limit.endpoint_override = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.387149] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_limit.insecure = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.387217] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_limit.keyfile = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.387388] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_limit.max_version = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.387590] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_limit.min_version = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.387684] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_limit.region_name = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.387959] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_limit.retriable_status_codes = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.388069] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_limit.service_name = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.388262] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_limit.service_type = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.388480] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_limit.split_loggers = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.388570] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_limit.status_code_retries = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.388734] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_limit.status_code_retry_delay = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.388941] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_limit.timeout = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.388995] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_limit.valid_interfaces = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.389190] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_limit.version = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.389394] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_reports.file_event_handler = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.389570] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_reports.file_event_handler_interval = 1 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.389658] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] oslo_reports.log_dir = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.389823] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.390051] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vif_plug_linux_bridge_privileged.group = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.390154] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.390386] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.394019] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.394019] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vif_plug_linux_bridge_privileged.user = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.394019] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.394019] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vif_plug_ovs_privileged.group = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.394019] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vif_plug_ovs_privileged.helper_command = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.394019] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.394019] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.394269] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] vif_plug_ovs_privileged.user = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.394269] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] os_vif_linux_bridge.flat_interface = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.394269] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.394269] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.394269] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.394269] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.394269] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.394447] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.394447] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] os_vif_linux_bridge.vlan_interface = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.394447] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.394447] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] os_vif_ovs.isolate_vif = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.394447] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.394447] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.394447] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.394654] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] os_vif_ovs.ovsdb_interface = native {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.394654] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] os_vif_ovs.per_port_bridge = False {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.394654] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] privsep_osbrick.capabilities = [21] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.394654] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] privsep_osbrick.group = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.394654] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] privsep_osbrick.helper_command = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.394773] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.394985] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] privsep_osbrick.thread_pool_size = 8 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.395033] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] privsep_osbrick.user = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.395173] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.395328] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] nova_sys_admin.group = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.395478] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] nova_sys_admin.helper_command = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.395635] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.395792] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] nova_sys_admin.thread_pool_size = 8 {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.395946] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] nova_sys_admin.user = None {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.396082] env[62460]: DEBUG oslo_service.service [None req-b657ae45-5c33-4675-acca-af82fbe85bfc None None] ******************************************************************************** {{(pid=62460) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 471.396571] env[62460]: INFO nova.service [-] Starting compute node (version 0.0.1) [ 471.900803] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-6ed0b664-2d20-40b2-8e10-e5af4b2a3d3a None None] Getting list of instances from cluster (obj){ [ 471.900803] env[62460]: value = "domain-c8" [ 471.900803] env[62460]: _type = "ClusterComputeResource" [ 471.900803] env[62460]: } {{(pid=62460) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 471.902045] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cea81f8-914a-4327-b944-4979048db290 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 471.910887] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-6ed0b664-2d20-40b2-8e10-e5af4b2a3d3a None None] Got total of 0 instances {{(pid=62460) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 471.911439] env[62460]: WARNING nova.virt.vmwareapi.driver [None req-6ed0b664-2d20-40b2-8e10-e5af4b2a3d3a None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 471.911896] env[62460]: INFO nova.virt.node [None req-6ed0b664-2d20-40b2-8e10-e5af4b2a3d3a None None] Generated node identity 76e8f00d-65cc-4766-bb9d-5006fb51c728 [ 471.912150] env[62460]: INFO nova.virt.node [None req-6ed0b664-2d20-40b2-8e10-e5af4b2a3d3a None None] Wrote node identity 76e8f00d-65cc-4766-bb9d-5006fb51c728 to /opt/stack/data/n-cpu-1/compute_id [ 472.414592] env[62460]: WARNING nova.compute.manager [None req-6ed0b664-2d20-40b2-8e10-e5af4b2a3d3a None None] Compute nodes ['76e8f00d-65cc-4766-bb9d-5006fb51c728'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 473.420872] env[62460]: INFO nova.compute.manager [None req-6ed0b664-2d20-40b2-8e10-e5af4b2a3d3a None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 474.427438] env[62460]: WARNING nova.compute.manager [None req-6ed0b664-2d20-40b2-8e10-e5af4b2a3d3a None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 474.427783] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6ed0b664-2d20-40b2-8e10-e5af4b2a3d3a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 474.427783] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6ed0b664-2d20-40b2-8e10-e5af4b2a3d3a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 474.427920] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6ed0b664-2d20-40b2-8e10-e5af4b2a3d3a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 474.428098] env[62460]: DEBUG nova.compute.resource_tracker [None req-6ed0b664-2d20-40b2-8e10-e5af4b2a3d3a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62460) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 474.429016] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a868f8f-f0aa-4fa8-ab17-7abab59cb605 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.437256] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c685e396-9b9d-404a-8787-0bc82abfdf5f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.450569] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d730ae0d-41f6-4e67-be7d-25f71a724f4d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.457131] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5ddecff-9f9c-479d-bc89-9b249b50ef00 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.486029] env[62460]: DEBUG nova.compute.resource_tracker [None req-6ed0b664-2d20-40b2-8e10-e5af4b2a3d3a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181494MB free_disk=135GB free_vcpus=48 pci_devices=None {{(pid=62460) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 474.486214] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6ed0b664-2d20-40b2-8e10-e5af4b2a3d3a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 474.486388] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6ed0b664-2d20-40b2-8e10-e5af4b2a3d3a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 474.992106] env[62460]: WARNING nova.compute.resource_tracker [None req-6ed0b664-2d20-40b2-8e10-e5af4b2a3d3a None None] No compute node record for cpu-1:76e8f00d-65cc-4766-bb9d-5006fb51c728: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host 76e8f00d-65cc-4766-bb9d-5006fb51c728 could not be found. [ 475.495867] env[62460]: INFO nova.compute.resource_tracker [None req-6ed0b664-2d20-40b2-8e10-e5af4b2a3d3a None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: 76e8f00d-65cc-4766-bb9d-5006fb51c728 [ 477.003418] env[62460]: DEBUG nova.compute.resource_tracker [None req-6ed0b664-2d20-40b2-8e10-e5af4b2a3d3a None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=62460) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 477.003839] env[62460]: DEBUG nova.compute.resource_tracker [None req-6ed0b664-2d20-40b2-8e10-e5af4b2a3d3a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=62460) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 477.156048] env[62460]: INFO nova.scheduler.client.report [None req-6ed0b664-2d20-40b2-8e10-e5af4b2a3d3a None None] [req-f222ff3a-8343-4c1b-84b6-ba37b0c1360a] Created resource provider record via placement API for resource provider with UUID 76e8f00d-65cc-4766-bb9d-5006fb51c728 and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 477.171929] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b49ff3ca-35cf-4e58-a5b4-f77bf871a665 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 477.179642] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cec597e-df63-43ee-92a7-4388222da3f7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 477.208618] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d03159ab-9a67-4cbb-8f62-46818a850af1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 477.215479] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7b457cf-81d1-403e-9818-90082ce4019a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 477.228030] env[62460]: DEBUG nova.compute.provider_tree [None req-6ed0b664-2d20-40b2-8e10-e5af4b2a3d3a None None] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 477.765668] env[62460]: DEBUG nova.scheduler.client.report [None req-6ed0b664-2d20-40b2-8e10-e5af4b2a3d3a None None] Updated inventory for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 477.765914] env[62460]: DEBUG nova.compute.provider_tree [None req-6ed0b664-2d20-40b2-8e10-e5af4b2a3d3a None None] Updating resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 generation from 0 to 1 during operation: update_inventory {{(pid=62460) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 477.766080] env[62460]: DEBUG nova.compute.provider_tree [None req-6ed0b664-2d20-40b2-8e10-e5af4b2a3d3a None None] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 477.817142] env[62460]: DEBUG nova.compute.provider_tree [None req-6ed0b664-2d20-40b2-8e10-e5af4b2a3d3a None None] Updating resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 generation from 1 to 2 during operation: update_traits {{(pid=62460) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 478.321790] env[62460]: DEBUG nova.compute.resource_tracker [None req-6ed0b664-2d20-40b2-8e10-e5af4b2a3d3a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62460) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 478.322201] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6ed0b664-2d20-40b2-8e10-e5af4b2a3d3a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.836s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 478.322201] env[62460]: DEBUG nova.service [None req-6ed0b664-2d20-40b2-8e10-e5af4b2a3d3a None None] Creating RPC server for service compute {{(pid=62460) start /opt/stack/nova/nova/service.py:186}} [ 478.339069] env[62460]: DEBUG nova.service [None req-6ed0b664-2d20-40b2-8e10-e5af4b2a3d3a None None] Join ServiceGroup membership for this service compute {{(pid=62460) start /opt/stack/nova/nova/service.py:203}} [ 478.339249] env[62460]: DEBUG nova.servicegroup.drivers.db [None req-6ed0b664-2d20-40b2-8e10-e5af4b2a3d3a None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=62460) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 480.341846] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._sync_power_states {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 480.845238] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Getting list of instances from cluster (obj){ [ 480.845238] env[62460]: value = "domain-c8" [ 480.845238] env[62460]: _type = "ClusterComputeResource" [ 480.845238] env[62460]: } {{(pid=62460) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 480.846450] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b67cb0f4-9f7d-440b-a3a7-c8db13d3696d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 480.855447] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Got total of 0 instances {{(pid=62460) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 480.855682] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 480.856026] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Getting list of instances from cluster (obj){ [ 480.856026] env[62460]: value = "domain-c8" [ 480.856026] env[62460]: _type = "ClusterComputeResource" [ 480.856026] env[62460]: } {{(pid=62460) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 480.856853] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90752aa0-925d-4193-ab73-624d4c282d17 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 480.864277] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Got total of 0 instances {{(pid=62460) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 516.568358] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Acquiring lock "44a2ac83-3fff-4958-a1d3-b3884b42c211" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 516.570121] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Lock "44a2ac83-3fff-4958-a1d3-b3884b42c211" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 517.072365] env[62460]: DEBUG nova.compute.manager [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 517.305460] env[62460]: DEBUG oslo_concurrency.lockutils [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] Acquiring lock "82afa576-f501-4a30-b8f7-e8b6334130cb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 517.305460] env[62460]: DEBUG oslo_concurrency.lockutils [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] Lock "82afa576-f501-4a30-b8f7-e8b6334130cb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 517.518113] env[62460]: DEBUG oslo_concurrency.lockutils [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Acquiring lock "61d0d996-78ad-4b20-a7f4-b8a8f130d969" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 517.518349] env[62460]: DEBUG oslo_concurrency.lockutils [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Lock "61d0d996-78ad-4b20-a7f4-b8a8f130d969" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 517.615046] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 517.615964] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 517.617709] env[62460]: INFO nova.compute.claims [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 517.675048] env[62460]: DEBUG oslo_concurrency.lockutils [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] Acquiring lock "7da5a822-6681-4dc6-b637-926ce425a042" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 517.675600] env[62460]: DEBUG oslo_concurrency.lockutils [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] Lock "7da5a822-6681-4dc6-b637-926ce425a042" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 517.807388] env[62460]: DEBUG nova.compute.manager [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 518.021749] env[62460]: DEBUG nova.compute.manager [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 518.183904] env[62460]: DEBUG nova.compute.manager [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] [instance: 7da5a822-6681-4dc6-b637-926ce425a042] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 518.342543] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Acquiring lock "2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 518.343649] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Lock "2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 518.346250] env[62460]: DEBUG oslo_concurrency.lockutils [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 518.555317] env[62460]: DEBUG oslo_concurrency.lockutils [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 518.718035] env[62460]: DEBUG oslo_concurrency.lockutils [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 518.734732] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d1cb7cc-7bb1-4d62-9e02-99e31edc978f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.743825] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc2f655f-c120-48cd-9e7a-bc6d321fae97 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.780410] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd626c68-dcd8-43ca-9f33-abed20fe8504 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.790164] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3108a240-2ccf-4e75-b797-8b847de22533 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.802788] env[62460]: DEBUG nova.compute.provider_tree [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 518.850590] env[62460]: DEBUG nova.compute.manager [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 519.310352] env[62460]: DEBUG nova.scheduler.client.report [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 519.383749] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 519.819030] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.200s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 519.819030] env[62460]: DEBUG nova.compute.manager [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 519.821913] env[62460]: DEBUG oslo_concurrency.lockutils [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.476s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 519.823653] env[62460]: INFO nova.compute.claims [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 520.331247] env[62460]: DEBUG nova.compute.utils [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 520.332442] env[62460]: DEBUG nova.compute.manager [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Not allocating networking since 'none' was specified. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 520.834415] env[62460]: DEBUG nova.compute.manager [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 520.977791] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22d56a1e-052c-4f08-9155-6f17d4bf35d4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.990266] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Acquiring lock "073b92a2-8db9-4337-af2e-5e85c78d4b93" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 520.990553] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Lock "073b92a2-8db9-4337-af2e-5e85c78d4b93" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 520.992668] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e5362f2-32d9-497e-8335-141bfe9d2642 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.033416] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6e40169-0cc0-479f-95e7-8e078204ad90 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.049423] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c223798-df26-492f-bfeb-8d803c6716b7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.063522] env[62460]: DEBUG nova.compute.provider_tree [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 521.498184] env[62460]: DEBUG nova.compute.manager [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 521.570491] env[62460]: DEBUG nova.scheduler.client.report [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 521.859645] env[62460]: DEBUG nova.compute.manager [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 522.039344] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 522.085896] env[62460]: DEBUG oslo_concurrency.lockutils [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.262s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 522.085896] env[62460]: DEBUG nova.compute.manager [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 522.086434] env[62460]: DEBUG oslo_concurrency.lockutils [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.531s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 522.090158] env[62460]: INFO nova.compute.claims [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 522.562323] env[62460]: DEBUG nova.virt.hardware [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 522.562794] env[62460]: DEBUG nova.virt.hardware [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 522.563015] env[62460]: DEBUG nova.virt.hardware [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 522.563695] env[62460]: DEBUG nova.virt.hardware [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 522.563884] env[62460]: DEBUG nova.virt.hardware [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 522.564083] env[62460]: DEBUG nova.virt.hardware [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 522.564364] env[62460]: DEBUG nova.virt.hardware [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 522.564918] env[62460]: DEBUG nova.virt.hardware [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 522.565415] env[62460]: DEBUG nova.virt.hardware [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 522.566184] env[62460]: DEBUG nova.virt.hardware [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 522.566440] env[62460]: DEBUG nova.virt.hardware [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 522.572561] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac73c660-4e8a-49a5-80af-7d505ca5ad43 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.583701] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1f25c9f-0cfb-4753-9ab5-a35e68cad3d5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.602555] env[62460]: DEBUG nova.compute.utils [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 522.609487] env[62460]: DEBUG nova.compute.manager [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 522.610044] env[62460]: DEBUG nova.network.neutron [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 522.619922] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-515c8d0f-1082-4bb5-9327-3432247447a2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.641494] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Instance VIF info [] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 522.652697] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 522.653440] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-631bb469-78ad-4a05-ab31-f3ea413b143e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.672141] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Created folder: OpenStack in parent group-v4. [ 522.672141] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Creating folder: Project (359d6c75b70140b7a88f6d6876461d5b). Parent ref: group-v281134. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 522.672141] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-785b74b0-0fcd-47c6-8cc6-efeb5c9b509d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.685702] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Created folder: Project (359d6c75b70140b7a88f6d6876461d5b) in parent group-v281134. [ 522.685895] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Creating folder: Instances. Parent ref: group-v281135. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 522.686149] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-50b2a3a5-6384-41bf-a3cb-d600020c0eab {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.697206] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Created folder: Instances in parent group-v281135. [ 522.697206] env[62460]: DEBUG oslo.service.loopingcall [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 522.697206] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 522.697206] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1ed5a3fc-41bd-4227-b2b7-cb3beeb9a327 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.719354] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 522.719354] env[62460]: value = "task-1313380" [ 522.719354] env[62460]: _type = "Task" [ 522.719354] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 522.727669] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313380, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 522.786762] env[62460]: DEBUG nova.policy [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f7320598c2564327b77f4f75a4dd1552', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd7bd664872b54cb0a05fc5ba3a6d35ad', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 523.110388] env[62460]: DEBUG nova.compute.manager [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 523.236423] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313380, 'name': CreateVM_Task, 'duration_secs': 0.350212} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 523.236423] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 523.237310] env[62460]: DEBUG oslo_vmware.service [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d8aced5-03c3-41d4-bb32-ff3894a78744 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.247974] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 523.247974] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 523.248662] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 523.249456] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f3360e9f-426d-471a-bc1a-b84aaaa02ae3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.259348] env[62460]: DEBUG oslo_vmware.api [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Waiting for the task: (returnval){ [ 523.259348] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]5234795f-1f10-85c4-36b6-4a9f007cc197" [ 523.259348] env[62460]: _type = "Task" [ 523.259348] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 523.275059] env[62460]: DEBUG oslo_vmware.api [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5234795f-1f10-85c4-36b6-4a9f007cc197, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 523.275059] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d0deba3-ce32-40ce-b0ce-fed26ff4eac4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.282596] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f35b9da-6f3d-46ab-a055-95c1d3b7e80c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.322903] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a578ec1-f7fc-407e-96cb-b4af2eec2931 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.331716] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-378b0d8e-b50d-4090-ab53-c1ed0299ef2f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.345596] env[62460]: DEBUG nova.compute.provider_tree [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 523.592514] env[62460]: DEBUG nova.network.neutron [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] Successfully created port: 1f7f19cc-6d5a-435a-8e7f-124093a4d283 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 523.773973] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 523.774215] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 523.774421] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 523.774585] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 523.775085] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 523.775227] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-24fdd043-f0f5-4970-a7b8-630263311fe0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.794253] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 523.794454] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 523.795568] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c0afbf4-eadc-47fe-9c5e-4fcabe083d9b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.805047] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-13b5316c-0798-401b-90fa-0e5624496f38 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.810736] env[62460]: DEBUG oslo_vmware.api [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Waiting for the task: (returnval){ [ 523.810736] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]5293688f-0dda-c5af-629f-8ca60f3860be" [ 523.810736] env[62460]: _type = "Task" [ 523.810736] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 523.818139] env[62460]: DEBUG oslo_vmware.api [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5293688f-0dda-c5af-629f-8ca60f3860be, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 523.848812] env[62460]: DEBUG nova.scheduler.client.report [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 524.124628] env[62460]: DEBUG nova.compute.manager [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 524.157490] env[62460]: DEBUG nova.virt.hardware [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 524.157743] env[62460]: DEBUG nova.virt.hardware [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 524.157904] env[62460]: DEBUG nova.virt.hardware [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 524.158098] env[62460]: DEBUG nova.virt.hardware [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 524.158499] env[62460]: DEBUG nova.virt.hardware [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 524.158733] env[62460]: DEBUG nova.virt.hardware [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 524.158910] env[62460]: DEBUG nova.virt.hardware [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 524.159220] env[62460]: DEBUG nova.virt.hardware [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 524.159278] env[62460]: DEBUG nova.virt.hardware [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 524.159475] env[62460]: DEBUG nova.virt.hardware [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 524.159623] env[62460]: DEBUG nova.virt.hardware [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 524.160682] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b70eb91-5a67-4610-be47-0776641871f6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.170702] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-370f0116-daf7-4c87-9341-5f81a26cb1d8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.324454] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Preparing fetch location {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 524.324454] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Creating directory with path [datastore1] vmware_temp/9871fda0-d4d1-4ffd-874e-26c4a0528c2f/eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 524.324854] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-00f6287b-fc4a-410e-90a0-d1d285264335 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.346282] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Created directory with path [datastore1] vmware_temp/9871fda0-d4d1-4ffd-874e-26c4a0528c2f/eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 524.346609] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Fetch image to [datastore1] vmware_temp/9871fda0-d4d1-4ffd-874e-26c4a0528c2f/eb3fb456-2dfd-4d95-978f-f17699d7735d/tmp-sparse.vmdk {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 524.347074] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Downloading image file data eb3fb456-2dfd-4d95-978f-f17699d7735d to [datastore1] vmware_temp/9871fda0-d4d1-4ffd-874e-26c4a0528c2f/eb3fb456-2dfd-4d95-978f-f17699d7735d/tmp-sparse.vmdk on the data store datastore1 {{(pid=62460) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 524.347791] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63f789a0-9a30-4228-b8c4-7e97fe548820 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.355915] env[62460]: DEBUG oslo_concurrency.lockutils [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.269s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 524.356535] env[62460]: DEBUG nova.compute.manager [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 524.359957] env[62460]: DEBUG oslo_concurrency.lockutils [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.642s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 524.362442] env[62460]: INFO nova.compute.claims [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] [instance: 7da5a822-6681-4dc6-b637-926ce425a042] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 524.365663] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19aa2849-b0f4-46ad-9b84-27447f5380ff {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.378990] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7c17799-2fd1-4481-a30b-5bfe114f468d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.420112] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2a6faca-d386-4c8e-825b-d433e876099b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.427540] env[62460]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-55290ae0-ca1d-4be6-b640-bc6d053ec07e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.514948] env[62460]: DEBUG nova.virt.vmwareapi.images [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Downloading image file data eb3fb456-2dfd-4d95-978f-f17699d7735d to the data store datastore1 {{(pid=62460) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 524.578373] env[62460]: DEBUG oslo_vmware.rw_handles [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/9871fda0-d4d1-4ffd-874e-26c4a0528c2f/eb3fb456-2dfd-4d95-978f-f17699d7735d/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=62460) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 524.873308] env[62460]: DEBUG nova.compute.utils [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 524.877451] env[62460]: DEBUG nova.compute.manager [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 524.877690] env[62460]: DEBUG nova.network.neutron [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 525.171243] env[62460]: DEBUG nova.policy [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a3c29d0a06624ad98155128ead12239f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '98f7cf2e2bc9462498b3c774b901136e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 525.324833] env[62460]: DEBUG oslo_vmware.rw_handles [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Completed reading data from the image iterator. {{(pid=62460) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 525.325072] env[62460]: DEBUG oslo_vmware.rw_handles [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Closing write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/9871fda0-d4d1-4ffd-874e-26c4a0528c2f/eb3fb456-2dfd-4d95-978f-f17699d7735d/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=62460) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 525.379287] env[62460]: DEBUG nova.compute.manager [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 525.386751] env[62460]: DEBUG nova.virt.vmwareapi.images [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Downloaded image file data eb3fb456-2dfd-4d95-978f-f17699d7735d to vmware_temp/9871fda0-d4d1-4ffd-874e-26c4a0528c2f/eb3fb456-2dfd-4d95-978f-f17699d7735d/tmp-sparse.vmdk on the data store datastore1 {{(pid=62460) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 525.390382] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Caching image {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 525.390611] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Copying Virtual Disk [datastore1] vmware_temp/9871fda0-d4d1-4ffd-874e-26c4a0528c2f/eb3fb456-2dfd-4d95-978f-f17699d7735d/tmp-sparse.vmdk to [datastore1] vmware_temp/9871fda0-d4d1-4ffd-874e-26c4a0528c2f/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 525.390883] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ae58c6cd-3226-49f5-8865-f249814f60cb {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.400285] env[62460]: DEBUG oslo_vmware.api [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Waiting for the task: (returnval){ [ 525.400285] env[62460]: value = "task-1313381" [ 525.400285] env[62460]: _type = "Task" [ 525.400285] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 525.412167] env[62460]: DEBUG oslo_vmware.api [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': task-1313381, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 525.529981] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da5047bd-6042-4914-a562-c25c6e939cda {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.544450] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2296e7fe-d9a0-4606-bdb5-e15782407f18 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.582533] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a20bd89-c96c-4971-bdf7-9ca5cdfe20df {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.590549] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74ee0d32-70e5-4ce9-8e8e-d954f0cf5f21 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.612510] env[62460]: DEBUG nova.compute.provider_tree [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 525.909579] env[62460]: DEBUG oslo_vmware.api [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': task-1313381, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 526.119309] env[62460]: DEBUG nova.scheduler.client.report [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 526.394302] env[62460]: DEBUG nova.compute.manager [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 526.412547] env[62460]: DEBUG oslo_vmware.api [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': task-1313381, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.680233} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 526.412863] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Copied Virtual Disk [datastore1] vmware_temp/9871fda0-d4d1-4ffd-874e-26c4a0528c2f/eb3fb456-2dfd-4d95-978f-f17699d7735d/tmp-sparse.vmdk to [datastore1] vmware_temp/9871fda0-d4d1-4ffd-874e-26c4a0528c2f/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 526.413095] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Deleting the datastore file [datastore1] vmware_temp/9871fda0-d4d1-4ffd-874e-26c4a0528c2f/eb3fb456-2dfd-4d95-978f-f17699d7735d/tmp-sparse.vmdk {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 526.413411] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-670c1dc3-aa0d-4ee2-8602-ec78c627c544 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.422725] env[62460]: DEBUG oslo_vmware.api [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Waiting for the task: (returnval){ [ 526.422725] env[62460]: value = "task-1313382" [ 526.422725] env[62460]: _type = "Task" [ 526.422725] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 526.431347] env[62460]: DEBUG oslo_vmware.api [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': task-1313382, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 526.433638] env[62460]: DEBUG nova.virt.hardware [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 526.433848] env[62460]: DEBUG nova.virt.hardware [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 526.433962] env[62460]: DEBUG nova.virt.hardware [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 526.434152] env[62460]: DEBUG nova.virt.hardware [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 526.434300] env[62460]: DEBUG nova.virt.hardware [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 526.434450] env[62460]: DEBUG nova.virt.hardware [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 526.434715] env[62460]: DEBUG nova.virt.hardware [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 526.435423] env[62460]: DEBUG nova.virt.hardware [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 526.435423] env[62460]: DEBUG nova.virt.hardware [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 526.435423] env[62460]: DEBUG nova.virt.hardware [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 526.435423] env[62460]: DEBUG nova.virt.hardware [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 526.436271] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ba23657-6900-4995-a27a-46c4215c61f9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.444238] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14d0be57-5c01-481a-ba5f-280762fbb2c9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.628240] env[62460]: DEBUG oslo_concurrency.lockutils [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.265s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 526.628240] env[62460]: DEBUG nova.compute.manager [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] [instance: 7da5a822-6681-4dc6-b637-926ce425a042] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 526.630717] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.247s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 526.633255] env[62460]: INFO nova.compute.claims [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 526.878027] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 526.878127] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 526.878353] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Starting heal instance info cache {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 526.878430] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Rebuilding the list of instances to heal {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 526.882715] env[62460]: DEBUG nova.network.neutron [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] Successfully created port: e30693e4-484a-4acc-ab76-e9da7cda9a03 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 526.935690] env[62460]: DEBUG oslo_vmware.api [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': task-1313382, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.024632} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 526.936182] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 526.936585] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Moving file from [datastore1] vmware_temp/9871fda0-d4d1-4ffd-874e-26c4a0528c2f/eb3fb456-2dfd-4d95-978f-f17699d7735d to [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d. {{(pid=62460) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 526.936981] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-c0f6983c-8477-41da-a246-c7ff653803e8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.946292] env[62460]: DEBUG oslo_vmware.api [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Waiting for the task: (returnval){ [ 526.946292] env[62460]: value = "task-1313383" [ 526.946292] env[62460]: _type = "Task" [ 526.946292] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 526.956580] env[62460]: DEBUG oslo_vmware.api [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': task-1313383, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 527.133233] env[62460]: DEBUG nova.compute.utils [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 527.134675] env[62460]: DEBUG nova.compute.manager [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] [instance: 7da5a822-6681-4dc6-b637-926ce425a042] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 527.134985] env[62460]: DEBUG nova.network.neutron [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] [instance: 7da5a822-6681-4dc6-b637-926ce425a042] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 527.346566] env[62460]: DEBUG nova.policy [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '49a7d86f7d1d4fc08a0541f030994a6f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c60dbb576cb3445dadf79d1840d1039b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 527.386461] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Skipping network cache update for instance because it is Building. {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 527.386644] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] Skipping network cache update for instance because it is Building. {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 527.386780] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] Skipping network cache update for instance because it is Building. {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 527.386908] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 7da5a822-6681-4dc6-b637-926ce425a042] Skipping network cache update for instance because it is Building. {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 527.387042] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] Skipping network cache update for instance because it is Building. {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 527.387988] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Didn't find any instances for network info cache update. {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 527.387988] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 527.387988] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 527.387988] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 527.387988] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 527.390120] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 527.390529] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 527.391423] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62460) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 527.391791] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 527.459485] env[62460]: DEBUG oslo_vmware.api [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': task-1313383, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.023255} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 527.459764] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] File moved {{(pid=62460) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 527.461014] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Cleaning up location [datastore1] vmware_temp/9871fda0-d4d1-4ffd-874e-26c4a0528c2f {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 527.461014] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Deleting the datastore file [datastore1] vmware_temp/9871fda0-d4d1-4ffd-874e-26c4a0528c2f {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 527.461014] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7bd8cc49-a200-431f-b726-d08f3447aec7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.470590] env[62460]: DEBUG oslo_vmware.api [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Waiting for the task: (returnval){ [ 527.470590] env[62460]: value = "task-1313384" [ 527.470590] env[62460]: _type = "Task" [ 527.470590] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 527.481230] env[62460]: DEBUG oslo_vmware.api [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': task-1313384, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 527.641490] env[62460]: DEBUG nova.compute.manager [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] [instance: 7da5a822-6681-4dc6-b637-926ce425a042] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 527.779952] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d5673a8-7ac6-45b8-affa-1b3d4f463fad {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.789076] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efcfc1a4-024e-436d-a87a-cf0df1351c1a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.824217] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12b1d936-fb4f-4f1a-aa41-bf8ca6388e36 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.831837] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-293a187a-3d4a-4838-98ea-bf625e932a62 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.845647] env[62460]: DEBUG nova.compute.provider_tree [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 527.896606] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 527.984439] env[62460]: DEBUG oslo_vmware.api [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': task-1313384, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.037102} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 527.985034] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 527.985464] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9130299f-2d5a-42a2-b690-01502ef856c1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.992137] env[62460]: DEBUG oslo_vmware.api [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Waiting for the task: (returnval){ [ 527.992137] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]5249be2d-81c4-8411-14f4-8ad0bb4723f1" [ 527.992137] env[62460]: _type = "Task" [ 527.992137] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 528.001714] env[62460]: DEBUG oslo_vmware.api [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5249be2d-81c4-8411-14f4-8ad0bb4723f1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 528.350754] env[62460]: DEBUG nova.scheduler.client.report [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 528.508710] env[62460]: DEBUG oslo_vmware.api [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5249be2d-81c4-8411-14f4-8ad0bb4723f1, 'name': SearchDatastore_Task, 'duration_secs': 0.011415} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 528.509079] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 528.511034] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 44a2ac83-3fff-4958-a1d3-b3884b42c211/44a2ac83-3fff-4958-a1d3-b3884b42c211.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 528.511034] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6db85561-6361-489a-ab0c-aa2095128559 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.518180] env[62460]: DEBUG oslo_vmware.api [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Waiting for the task: (returnval){ [ 528.518180] env[62460]: value = "task-1313385" [ 528.518180] env[62460]: _type = "Task" [ 528.518180] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 528.527889] env[62460]: DEBUG oslo_vmware.api [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': task-1313385, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 528.655539] env[62460]: DEBUG nova.compute.manager [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] [instance: 7da5a822-6681-4dc6-b637-926ce425a042] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 528.696704] env[62460]: DEBUG nova.virt.hardware [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 528.696935] env[62460]: DEBUG nova.virt.hardware [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 528.699644] env[62460]: DEBUG nova.virt.hardware [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 528.701791] env[62460]: DEBUG nova.virt.hardware [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 528.701791] env[62460]: DEBUG nova.virt.hardware [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 528.701791] env[62460]: DEBUG nova.virt.hardware [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 528.701791] env[62460]: DEBUG nova.virt.hardware [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 528.701791] env[62460]: DEBUG nova.virt.hardware [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 528.702131] env[62460]: DEBUG nova.virt.hardware [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 528.702131] env[62460]: DEBUG nova.virt.hardware [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 528.702131] env[62460]: DEBUG nova.virt.hardware [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 528.702966] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-950a6a4c-24f2-489e-9704-25669b553432 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.711906] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba9f4669-e9d0-4f5e-9634-bc738056752a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.787697] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Acquiring lock "48e50ab0-9d4e-4c5d-96d5-55399b3a8e19" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 528.788172] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Lock "48e50ab0-9d4e-4c5d-96d5-55399b3a8e19" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 528.857686] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.227s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 528.859226] env[62460]: DEBUG nova.compute.manager [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 528.864486] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.824s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 528.866065] env[62460]: INFO nova.compute.claims [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 529.029068] env[62460]: DEBUG oslo_vmware.api [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': task-1313385, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 529.296044] env[62460]: DEBUG nova.compute.manager [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 529.364411] env[62460]: DEBUG nova.compute.utils [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 529.367894] env[62460]: DEBUG nova.compute.manager [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 529.367894] env[62460]: DEBUG nova.network.neutron [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 529.497292] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] Acquiring lock "d16440e7-aa8a-4041-a4cd-d0ea84d8c027" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 529.497509] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] Lock "d16440e7-aa8a-4041-a4cd-d0ea84d8c027" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 529.534563] env[62460]: DEBUG oslo_vmware.api [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': task-1313385, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.624563} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 529.534563] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 44a2ac83-3fff-4958-a1d3-b3884b42c211/44a2ac83-3fff-4958-a1d3-b3884b42c211.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 529.534862] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 529.535949] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1ae53156-f6a0-40f8-98c8-a84182b01283 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.547328] env[62460]: DEBUG oslo_vmware.api [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Waiting for the task: (returnval){ [ 529.547328] env[62460]: value = "task-1313386" [ 529.547328] env[62460]: _type = "Task" [ 529.547328] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 529.556764] env[62460]: DEBUG oslo_vmware.api [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': task-1313386, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 529.558753] env[62460]: DEBUG nova.policy [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8e09d51001bc4e40b4d35db1511d1a6e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '210ef0c43440478b9dcc5360edf52241', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 529.566531] env[62460]: DEBUG nova.network.neutron [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] [instance: 7da5a822-6681-4dc6-b637-926ce425a042] Successfully created port: d8646801-5db8-41d8-a53c-f24e8f20d19a {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 529.828782] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 529.871268] env[62460]: DEBUG nova.compute.manager [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 530.003398] env[62460]: DEBUG nova.compute.manager [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 530.078728] env[62460]: DEBUG oslo_vmware.api [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': task-1313386, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064644} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 530.078728] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 530.078728] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a553a5a-8d85-480c-91bf-ac269a242045 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.094585] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Reconfiguring VM instance instance-00000001 to attach disk [datastore1] 44a2ac83-3fff-4958-a1d3-b3884b42c211/44a2ac83-3fff-4958-a1d3-b3884b42c211.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 530.096042] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a6355783-0d13-455f-ba3a-ddf7f5500fc9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.130457] env[62460]: DEBUG oslo_vmware.api [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Waiting for the task: (returnval){ [ 530.130457] env[62460]: value = "task-1313387" [ 530.130457] env[62460]: _type = "Task" [ 530.130457] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 530.141513] env[62460]: DEBUG oslo_vmware.api [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': task-1313387, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 530.168313] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-316b5627-5866-4980-bbd0-1236e7130ccb {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.178546] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f844c62-1166-464f-bf98-c44c002e431a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.215340] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b82ee453-330b-4264-9f00-a1dac062db7d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.226176] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86267290-5ec8-4fd7-abad-8c5da8e5e6b4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.247543] env[62460]: DEBUG nova.compute.provider_tree [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 530.533598] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 530.629733] env[62460]: ERROR nova.compute.manager [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1f7f19cc-6d5a-435a-8e7f-124093a4d283, please check neutron logs for more information. [ 530.629733] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 530.629733] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 530.629733] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 530.629733] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 530.629733] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 530.629733] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 530.629733] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 530.629733] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 530.629733] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 530.629733] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 530.629733] env[62460]: ERROR nova.compute.manager raise self.value [ 530.629733] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 530.629733] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 530.629733] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 530.629733] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 530.630971] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 530.630971] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 530.630971] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1f7f19cc-6d5a-435a-8e7f-124093a4d283, please check neutron logs for more information. [ 530.630971] env[62460]: ERROR nova.compute.manager [ 530.630971] env[62460]: Traceback (most recent call last): [ 530.632264] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 530.632264] env[62460]: listener.cb(fileno) [ 530.632264] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 530.632264] env[62460]: result = function(*args, **kwargs) [ 530.632264] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 530.632264] env[62460]: return func(*args, **kwargs) [ 530.632264] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 530.632264] env[62460]: raise e [ 530.632264] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 530.632264] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 530.632264] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 530.632264] env[62460]: created_port_ids = self._update_ports_for_instance( [ 530.632264] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 530.632264] env[62460]: with excutils.save_and_reraise_exception(): [ 530.632264] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 530.632264] env[62460]: self.force_reraise() [ 530.632264] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 530.632264] env[62460]: raise self.value [ 530.632264] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 530.632264] env[62460]: updated_port = self._update_port( [ 530.632913] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 530.632913] env[62460]: _ensure_no_port_binding_failure(port) [ 530.632913] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 530.632913] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 530.632913] env[62460]: nova.exception.PortBindingFailed: Binding failed for port 1f7f19cc-6d5a-435a-8e7f-124093a4d283, please check neutron logs for more information. [ 530.632913] env[62460]: Removing descriptor: 15 [ 530.633960] env[62460]: ERROR nova.compute.manager [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1f7f19cc-6d5a-435a-8e7f-124093a4d283, please check neutron logs for more information. [ 530.633960] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] Traceback (most recent call last): [ 530.633960] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 530.633960] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] yield resources [ 530.633960] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 530.633960] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] self.driver.spawn(context, instance, image_meta, [ 530.633960] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 530.633960] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 530.633960] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 530.633960] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] vm_ref = self.build_virtual_machine(instance, [ 530.633960] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 530.634363] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] vif_infos = vmwarevif.get_vif_info(self._session, [ 530.634363] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 530.634363] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] for vif in network_info: [ 530.634363] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 530.634363] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] return self._sync_wrapper(fn, *args, **kwargs) [ 530.634363] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 530.634363] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] self.wait() [ 530.634363] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 530.634363] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] self[:] = self._gt.wait() [ 530.634363] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 530.634363] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] return self._exit_event.wait() [ 530.634363] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 530.634363] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] result = hub.switch() [ 530.635174] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 530.635174] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] return self.greenlet.switch() [ 530.635174] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 530.635174] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] result = function(*args, **kwargs) [ 530.635174] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 530.635174] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] return func(*args, **kwargs) [ 530.635174] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 530.635174] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] raise e [ 530.635174] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 530.635174] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] nwinfo = self.network_api.allocate_for_instance( [ 530.635174] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 530.635174] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] created_port_ids = self._update_ports_for_instance( [ 530.635174] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 530.636097] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] with excutils.save_and_reraise_exception(): [ 530.636097] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 530.636097] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] self.force_reraise() [ 530.636097] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 530.636097] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] raise self.value [ 530.636097] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 530.636097] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] updated_port = self._update_port( [ 530.636097] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 530.636097] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] _ensure_no_port_binding_failure(port) [ 530.636097] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 530.636097] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] raise exception.PortBindingFailed(port_id=port['id']) [ 530.636097] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] nova.exception.PortBindingFailed: Binding failed for port 1f7f19cc-6d5a-435a-8e7f-124093a4d283, please check neutron logs for more information. [ 530.636097] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] [ 530.637160] env[62460]: INFO nova.compute.manager [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] Terminating instance [ 530.641557] env[62460]: DEBUG oslo_concurrency.lockutils [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] Acquiring lock "refresh_cache-82afa576-f501-4a30-b8f7-e8b6334130cb" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 530.643442] env[62460]: DEBUG oslo_concurrency.lockutils [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] Acquired lock "refresh_cache-82afa576-f501-4a30-b8f7-e8b6334130cb" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 530.643442] env[62460]: DEBUG nova.network.neutron [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 530.649401] env[62460]: DEBUG oslo_vmware.api [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': task-1313387, 'name': ReconfigVM_Task, 'duration_secs': 0.278991} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 530.649680] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Reconfigured VM instance instance-00000001 to attach disk [datastore1] 44a2ac83-3fff-4958-a1d3-b3884b42c211/44a2ac83-3fff-4958-a1d3-b3884b42c211.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 530.650785] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0b9ab10c-cb89-4fe2-b89a-019e9726d17b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.659626] env[62460]: DEBUG oslo_vmware.api [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Waiting for the task: (returnval){ [ 530.659626] env[62460]: value = "task-1313388" [ 530.659626] env[62460]: _type = "Task" [ 530.659626] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 530.669447] env[62460]: DEBUG oslo_vmware.api [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': task-1313388, 'name': Rename_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 530.750947] env[62460]: DEBUG nova.scheduler.client.report [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 530.892599] env[62460]: DEBUG nova.compute.manager [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 530.922473] env[62460]: DEBUG nova.network.neutron [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] Successfully created port: 228ca897-3a3c-4e88-9838-451cdc94d456 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 530.929782] env[62460]: DEBUG nova.virt.hardware [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 530.931044] env[62460]: DEBUG nova.virt.hardware [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 530.931044] env[62460]: DEBUG nova.virt.hardware [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 530.931044] env[62460]: DEBUG nova.virt.hardware [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 530.931228] env[62460]: DEBUG nova.virt.hardware [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 530.931341] env[62460]: DEBUG nova.virt.hardware [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 530.931872] env[62460]: DEBUG nova.virt.hardware [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 530.931946] env[62460]: DEBUG nova.virt.hardware [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 530.932345] env[62460]: DEBUG nova.virt.hardware [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 530.932484] env[62460]: DEBUG nova.virt.hardware [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 530.932788] env[62460]: DEBUG nova.virt.hardware [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 530.937756] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22d0e214-81b2-4809-9660-54b52ff7b2a8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.946797] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-920e974f-83c3-47f2-b063-5660fc98762e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.182240] env[62460]: DEBUG oslo_vmware.api [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': task-1313388, 'name': Rename_Task, 'duration_secs': 0.139238} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 531.182567] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 531.182838] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6fb7dfc3-1cd1-4742-9b8c-c8d719d15035 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.190500] env[62460]: DEBUG oslo_vmware.api [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Waiting for the task: (returnval){ [ 531.190500] env[62460]: value = "task-1313389" [ 531.190500] env[62460]: _type = "Task" [ 531.190500] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 531.197839] env[62460]: DEBUG oslo_vmware.api [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': task-1313389, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 531.211957] env[62460]: DEBUG nova.network.neutron [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 531.261668] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.399s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 531.262264] env[62460]: DEBUG nova.compute.manager [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 531.266390] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 3.370s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 531.266576] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 531.266770] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62460) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 531.267107] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.439s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 531.268704] env[62460]: INFO nova.compute.claims [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 531.274479] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14a78087-52b8-4252-9919-35b15818f23b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.281720] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6c0f82e-aaf0-4127-bb7d-cb5922c7fbdb {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.303521] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8482fa2c-6b70-43f3-aeec-52c9e9bf1f2a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.308889] env[62460]: DEBUG nova.compute.manager [req-2fca86c6-ab5a-442c-80aa-64638138faa8 req-61ec36b8-628b-4bdd-999a-a981dd74c4ce service nova] [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] Received event network-changed-1f7f19cc-6d5a-435a-8e7f-124093a4d283 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 531.308889] env[62460]: DEBUG nova.compute.manager [req-2fca86c6-ab5a-442c-80aa-64638138faa8 req-61ec36b8-628b-4bdd-999a-a981dd74c4ce service nova] [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] Refreshing instance network info cache due to event network-changed-1f7f19cc-6d5a-435a-8e7f-124093a4d283. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 531.308889] env[62460]: DEBUG oslo_concurrency.lockutils [req-2fca86c6-ab5a-442c-80aa-64638138faa8 req-61ec36b8-628b-4bdd-999a-a981dd74c4ce service nova] Acquiring lock "refresh_cache-82afa576-f501-4a30-b8f7-e8b6334130cb" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 531.315404] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41099a00-3231-4c77-97e6-6e270c863f2d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.355738] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181500MB free_disk=135GB free_vcpus=48 pci_devices=None {{(pid=62460) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 531.355872] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 531.374196] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] Acquiring lock "72714a2e-a95d-495e-b163-f5168148fcdf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 531.374497] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] Lock "72714a2e-a95d-495e-b163-f5168148fcdf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 531.517365] env[62460]: DEBUG nova.network.neutron [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 531.703181] env[62460]: DEBUG oslo_vmware.api [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': task-1313389, 'name': PowerOnVM_Task, 'duration_secs': 0.463702} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 531.703560] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 531.704801] env[62460]: INFO nova.compute.manager [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Took 9.84 seconds to spawn the instance on the hypervisor. [ 531.706701] env[62460]: DEBUG nova.compute.manager [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 531.706701] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27c5daad-e03f-44b8-8a64-449377b36bec {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.776326] env[62460]: DEBUG nova.compute.utils [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 531.776326] env[62460]: DEBUG nova.compute.manager [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 531.776326] env[62460]: DEBUG nova.network.neutron [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 531.878586] env[62460]: DEBUG nova.compute.manager [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 532.023317] env[62460]: DEBUG oslo_concurrency.lockutils [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] Releasing lock "refresh_cache-82afa576-f501-4a30-b8f7-e8b6334130cb" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 532.023453] env[62460]: DEBUG nova.compute.manager [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 532.023870] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 532.026530] env[62460]: DEBUG oslo_concurrency.lockutils [req-2fca86c6-ab5a-442c-80aa-64638138faa8 req-61ec36b8-628b-4bdd-999a-a981dd74c4ce service nova] Acquired lock "refresh_cache-82afa576-f501-4a30-b8f7-e8b6334130cb" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 532.026530] env[62460]: DEBUG nova.network.neutron [req-2fca86c6-ab5a-442c-80aa-64638138faa8 req-61ec36b8-628b-4bdd-999a-a981dd74c4ce service nova] [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] Refreshing network info cache for port 1f7f19cc-6d5a-435a-8e7f-124093a4d283 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 532.026530] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-318a0e40-e0a6-4370-b715-1663400edd8a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.037633] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2839776a-6280-4a03-9189-fb581bd0c832 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.068348] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 82afa576-f501-4a30-b8f7-e8b6334130cb could not be found. [ 532.068450] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 532.068631] env[62460]: INFO nova.compute.manager [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] Took 0.05 seconds to destroy the instance on the hypervisor. [ 532.068876] env[62460]: DEBUG oslo.service.loopingcall [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 532.072077] env[62460]: DEBUG nova.compute.manager [-] [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 532.072077] env[62460]: DEBUG nova.network.neutron [-] [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 532.096838] env[62460]: DEBUG nova.policy [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4d95c2a24f0841de8990410e24f325e9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '11491a5610734d96bac82035c46bd5d1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 532.133279] env[62460]: DEBUG nova.network.neutron [-] [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 532.227853] env[62460]: INFO nova.compute.manager [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Took 14.65 seconds to build instance. [ 532.284656] env[62460]: DEBUG nova.compute.manager [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 532.401148] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 532.541602] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6af9fd5-2d49-410b-ba48-65bb7505877b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.554591] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-618ba2e2-b21a-4207-a18f-17189f12b87f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.588167] env[62460]: DEBUG nova.network.neutron [req-2fca86c6-ab5a-442c-80aa-64638138faa8 req-61ec36b8-628b-4bdd-999a-a981dd74c4ce service nova] [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 532.590509] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-871b8f4b-ec59-4e10-a544-75902bf43628 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.598374] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16e109db-594b-4ad4-82de-23348bc17835 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.612396] env[62460]: DEBUG nova.compute.provider_tree [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 532.636448] env[62460]: DEBUG nova.network.neutron [-] [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 532.731623] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7f802588-df97-4be4-b18f-9460471367f9 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Lock "44a2ac83-3fff-4958-a1d3-b3884b42c211" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.162s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 532.782813] env[62460]: DEBUG nova.network.neutron [req-2fca86c6-ab5a-442c-80aa-64638138faa8 req-61ec36b8-628b-4bdd-999a-a981dd74c4ce service nova] [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 533.121090] env[62460]: DEBUG nova.scheduler.client.report [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 533.141067] env[62460]: INFO nova.compute.manager [-] [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] Took 1.07 seconds to deallocate network for instance. [ 533.150372] env[62460]: DEBUG nova.compute.claims [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 533.150372] env[62460]: DEBUG oslo_concurrency.lockutils [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 533.285941] env[62460]: DEBUG oslo_concurrency.lockutils [req-2fca86c6-ab5a-442c-80aa-64638138faa8 req-61ec36b8-628b-4bdd-999a-a981dd74c4ce service nova] Releasing lock "refresh_cache-82afa576-f501-4a30-b8f7-e8b6334130cb" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 533.301339] env[62460]: DEBUG nova.compute.manager [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 533.338102] env[62460]: DEBUG nova.virt.hardware [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 533.338102] env[62460]: DEBUG nova.virt.hardware [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 533.338102] env[62460]: DEBUG nova.virt.hardware [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 533.338102] env[62460]: DEBUG nova.virt.hardware [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 533.338398] env[62460]: DEBUG nova.virt.hardware [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 533.338398] env[62460]: DEBUG nova.virt.hardware [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 533.338398] env[62460]: DEBUG nova.virt.hardware [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 533.338398] env[62460]: DEBUG nova.virt.hardware [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 533.338398] env[62460]: DEBUG nova.virt.hardware [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 533.339856] env[62460]: DEBUG nova.virt.hardware [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 533.339856] env[62460]: DEBUG nova.virt.hardware [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 533.340616] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e844e73a-2c87-4bef-9032-bd576b1e600a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.352986] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f09c63b-9688-41c4-a773-d99b961b9749 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.452519] env[62460]: ERROR nova.compute.manager [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e30693e4-484a-4acc-ab76-e9da7cda9a03, please check neutron logs for more information. [ 533.452519] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 533.452519] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 533.452519] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 533.452519] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 533.452519] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 533.452519] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 533.452519] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 533.452519] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 533.452519] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 533.452519] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 533.452519] env[62460]: ERROR nova.compute.manager raise self.value [ 533.452519] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 533.452519] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 533.452519] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 533.452519] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 533.453513] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 533.453513] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 533.453513] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e30693e4-484a-4acc-ab76-e9da7cda9a03, please check neutron logs for more information. [ 533.453513] env[62460]: ERROR nova.compute.manager [ 533.453513] env[62460]: Traceback (most recent call last): [ 533.453513] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 533.453513] env[62460]: listener.cb(fileno) [ 533.453513] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 533.453513] env[62460]: result = function(*args, **kwargs) [ 533.453513] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 533.453513] env[62460]: return func(*args, **kwargs) [ 533.453513] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 533.453513] env[62460]: raise e [ 533.453513] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 533.453513] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 533.453513] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 533.453513] env[62460]: created_port_ids = self._update_ports_for_instance( [ 533.453513] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 533.453513] env[62460]: with excutils.save_and_reraise_exception(): [ 533.453513] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 533.453513] env[62460]: self.force_reraise() [ 533.453513] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 533.453513] env[62460]: raise self.value [ 533.453513] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 533.453513] env[62460]: updated_port = self._update_port( [ 533.453513] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 533.453513] env[62460]: _ensure_no_port_binding_failure(port) [ 533.453513] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 533.453513] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 533.454763] env[62460]: nova.exception.PortBindingFailed: Binding failed for port e30693e4-484a-4acc-ab76-e9da7cda9a03, please check neutron logs for more information. [ 533.454763] env[62460]: Removing descriptor: 18 [ 533.454763] env[62460]: ERROR nova.compute.manager [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e30693e4-484a-4acc-ab76-e9da7cda9a03, please check neutron logs for more information. [ 533.454763] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] Traceback (most recent call last): [ 533.454763] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 533.454763] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] yield resources [ 533.454763] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 533.454763] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] self.driver.spawn(context, instance, image_meta, [ 533.454763] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 533.454763] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] self._vmops.spawn(context, instance, image_meta, injected_files, [ 533.454763] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 533.454763] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] vm_ref = self.build_virtual_machine(instance, [ 533.455272] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 533.455272] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] vif_infos = vmwarevif.get_vif_info(self._session, [ 533.455272] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 533.455272] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] for vif in network_info: [ 533.455272] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 533.455272] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] return self._sync_wrapper(fn, *args, **kwargs) [ 533.455272] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 533.455272] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] self.wait() [ 533.455272] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 533.455272] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] self[:] = self._gt.wait() [ 533.455272] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 533.455272] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] return self._exit_event.wait() [ 533.455272] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 533.455630] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] result = hub.switch() [ 533.455630] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 533.455630] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] return self.greenlet.switch() [ 533.455630] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 533.455630] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] result = function(*args, **kwargs) [ 533.455630] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 533.455630] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] return func(*args, **kwargs) [ 533.455630] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 533.455630] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] raise e [ 533.455630] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 533.455630] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] nwinfo = self.network_api.allocate_for_instance( [ 533.455630] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 533.455630] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] created_port_ids = self._update_ports_for_instance( [ 533.460301] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 533.460301] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] with excutils.save_and_reraise_exception(): [ 533.460301] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 533.460301] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] self.force_reraise() [ 533.460301] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 533.460301] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] raise self.value [ 533.460301] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 533.460301] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] updated_port = self._update_port( [ 533.460301] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 533.460301] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] _ensure_no_port_binding_failure(port) [ 533.460301] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 533.460301] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] raise exception.PortBindingFailed(port_id=port['id']) [ 533.460760] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] nova.exception.PortBindingFailed: Binding failed for port e30693e4-484a-4acc-ab76-e9da7cda9a03, please check neutron logs for more information. [ 533.460760] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] [ 533.460760] env[62460]: INFO nova.compute.manager [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] Terminating instance [ 533.460760] env[62460]: DEBUG oslo_concurrency.lockutils [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Acquiring lock "refresh_cache-61d0d996-78ad-4b20-a7f4-b8a8f130d969" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 533.460760] env[62460]: DEBUG oslo_concurrency.lockutils [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Acquired lock "refresh_cache-61d0d996-78ad-4b20-a7f4-b8a8f130d969" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 533.460760] env[62460]: DEBUG nova.network.neutron [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 533.627018] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.360s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 533.627588] env[62460]: DEBUG nova.compute.manager [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 533.630778] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.098s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 533.632368] env[62460]: INFO nova.compute.claims [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 533.864623] env[62460]: INFO nova.compute.manager [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Rebuilding instance [ 533.914663] env[62460]: DEBUG nova.compute.manager [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 533.916340] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-382b0bb6-0bfa-46ae-94e4-d7660388b6f3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.036908] env[62460]: DEBUG nova.network.neutron [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 534.140481] env[62460]: DEBUG nova.compute.utils [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 534.143894] env[62460]: DEBUG nova.compute.manager [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 534.144213] env[62460]: DEBUG nova.network.neutron [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 534.284498] env[62460]: DEBUG nova.network.neutron [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 534.289018] env[62460]: DEBUG nova.network.neutron [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] Successfully created port: 6d1179e6-551c-4986-b849-eceeefc23045 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 534.398788] env[62460]: ERROR nova.compute.manager [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 228ca897-3a3c-4e88-9838-451cdc94d456, please check neutron logs for more information. [ 534.398788] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 534.398788] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 534.398788] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 534.398788] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 534.398788] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 534.398788] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 534.398788] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 534.398788] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 534.398788] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 534.398788] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 534.398788] env[62460]: ERROR nova.compute.manager raise self.value [ 534.398788] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 534.398788] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 534.398788] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 534.398788] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 534.399236] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 534.399236] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 534.399236] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 228ca897-3a3c-4e88-9838-451cdc94d456, please check neutron logs for more information. [ 534.399236] env[62460]: ERROR nova.compute.manager [ 534.399236] env[62460]: Traceback (most recent call last): [ 534.399236] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 534.399236] env[62460]: listener.cb(fileno) [ 534.399236] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 534.399236] env[62460]: result = function(*args, **kwargs) [ 534.399236] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 534.399236] env[62460]: return func(*args, **kwargs) [ 534.399236] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 534.399236] env[62460]: raise e [ 534.399236] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 534.399236] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 534.399236] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 534.399236] env[62460]: created_port_ids = self._update_ports_for_instance( [ 534.399236] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 534.399236] env[62460]: with excutils.save_and_reraise_exception(): [ 534.399236] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 534.399236] env[62460]: self.force_reraise() [ 534.399236] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 534.399236] env[62460]: raise self.value [ 534.399236] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 534.399236] env[62460]: updated_port = self._update_port( [ 534.399236] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 534.399236] env[62460]: _ensure_no_port_binding_failure(port) [ 534.399236] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 534.399236] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 534.400135] env[62460]: nova.exception.PortBindingFailed: Binding failed for port 228ca897-3a3c-4e88-9838-451cdc94d456, please check neutron logs for more information. [ 534.400135] env[62460]: Removing descriptor: 17 [ 534.400135] env[62460]: ERROR nova.compute.manager [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 228ca897-3a3c-4e88-9838-451cdc94d456, please check neutron logs for more information. [ 534.400135] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] Traceback (most recent call last): [ 534.400135] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 534.400135] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] yield resources [ 534.400135] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 534.400135] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] self.driver.spawn(context, instance, image_meta, [ 534.400135] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 534.400135] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 534.400135] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 534.400135] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] vm_ref = self.build_virtual_machine(instance, [ 534.400468] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 534.400468] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] vif_infos = vmwarevif.get_vif_info(self._session, [ 534.400468] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 534.400468] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] for vif in network_info: [ 534.400468] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 534.400468] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] return self._sync_wrapper(fn, *args, **kwargs) [ 534.400468] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 534.400468] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] self.wait() [ 534.400468] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 534.400468] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] self[:] = self._gt.wait() [ 534.400468] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 534.400468] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] return self._exit_event.wait() [ 534.400468] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 534.400805] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] result = hub.switch() [ 534.400805] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 534.400805] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] return self.greenlet.switch() [ 534.400805] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 534.400805] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] result = function(*args, **kwargs) [ 534.400805] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 534.400805] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] return func(*args, **kwargs) [ 534.400805] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 534.400805] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] raise e [ 534.400805] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 534.400805] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] nwinfo = self.network_api.allocate_for_instance( [ 534.400805] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 534.400805] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] created_port_ids = self._update_ports_for_instance( [ 534.401130] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 534.401130] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] with excutils.save_and_reraise_exception(): [ 534.401130] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 534.401130] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] self.force_reraise() [ 534.401130] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 534.401130] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] raise self.value [ 534.401130] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 534.401130] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] updated_port = self._update_port( [ 534.401130] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 534.401130] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] _ensure_no_port_binding_failure(port) [ 534.401130] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 534.401130] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] raise exception.PortBindingFailed(port_id=port['id']) [ 534.401877] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] nova.exception.PortBindingFailed: Binding failed for port 228ca897-3a3c-4e88-9838-451cdc94d456, please check neutron logs for more information. [ 534.401877] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] [ 534.401877] env[62460]: INFO nova.compute.manager [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] Terminating instance [ 534.406523] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Acquiring lock "refresh_cache-2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 534.406770] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Acquired lock "refresh_cache-2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 534.406956] env[62460]: DEBUG nova.network.neutron [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 534.432271] env[62460]: DEBUG nova.policy [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5d0676367f10488e9aa21fa9f64bbb2f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '91a65d30ffe14c9f8c5d869164ce7b86', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 534.434264] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 534.434585] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4ea3d60a-9107-4c23-af60-07b332fabce1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.444772] env[62460]: DEBUG oslo_vmware.api [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Waiting for the task: (returnval){ [ 534.444772] env[62460]: value = "task-1313394" [ 534.444772] env[62460]: _type = "Task" [ 534.444772] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 534.455049] env[62460]: DEBUG oslo_vmware.api [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': task-1313394, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 534.598203] env[62460]: DEBUG nova.compute.manager [req-f7146609-282a-4a2a-b2a5-9d4843aec179 req-fc9e11cd-b0c6-495a-9ba7-124ea14a6efe service nova] [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] Received event network-vif-deleted-1f7f19cc-6d5a-435a-8e7f-124093a4d283 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 534.598392] env[62460]: DEBUG nova.compute.manager [req-f7146609-282a-4a2a-b2a5-9d4843aec179 req-fc9e11cd-b0c6-495a-9ba7-124ea14a6efe service nova] [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] Received event network-changed-e30693e4-484a-4acc-ab76-e9da7cda9a03 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 534.598654] env[62460]: DEBUG nova.compute.manager [req-f7146609-282a-4a2a-b2a5-9d4843aec179 req-fc9e11cd-b0c6-495a-9ba7-124ea14a6efe service nova] [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] Refreshing instance network info cache due to event network-changed-e30693e4-484a-4acc-ab76-e9da7cda9a03. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 534.599018] env[62460]: DEBUG oslo_concurrency.lockutils [req-f7146609-282a-4a2a-b2a5-9d4843aec179 req-fc9e11cd-b0c6-495a-9ba7-124ea14a6efe service nova] Acquiring lock "refresh_cache-61d0d996-78ad-4b20-a7f4-b8a8f130d969" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 534.653358] env[62460]: DEBUG nova.compute.manager [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 534.787480] env[62460]: DEBUG oslo_concurrency.lockutils [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Releasing lock "refresh_cache-61d0d996-78ad-4b20-a7f4-b8a8f130d969" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 534.789417] env[62460]: DEBUG nova.compute.manager [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 534.789644] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 534.790140] env[62460]: DEBUG oslo_concurrency.lockutils [req-f7146609-282a-4a2a-b2a5-9d4843aec179 req-fc9e11cd-b0c6-495a-9ba7-124ea14a6efe service nova] Acquired lock "refresh_cache-61d0d996-78ad-4b20-a7f4-b8a8f130d969" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 534.790595] env[62460]: DEBUG nova.network.neutron [req-f7146609-282a-4a2a-b2a5-9d4843aec179 req-fc9e11cd-b0c6-495a-9ba7-124ea14a6efe service nova] [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] Refreshing network info cache for port e30693e4-484a-4acc-ab76-e9da7cda9a03 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 534.791555] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-af371025-9f0f-4a40-84db-a21c6b880279 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.812181] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b196c505-6a23-4d01-b400-d2763f2b293d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.845056] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 61d0d996-78ad-4b20-a7f4-b8a8f130d969 could not be found. [ 534.845056] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 534.845056] env[62460]: INFO nova.compute.manager [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] Took 0.05 seconds to destroy the instance on the hypervisor. [ 534.845056] env[62460]: DEBUG oslo.service.loopingcall [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 534.845056] env[62460]: DEBUG nova.compute.manager [-] [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 534.845056] env[62460]: DEBUG nova.network.neutron [-] [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 534.885375] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f572157d-33ad-4077-9511-013c6210c7ce {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.894192] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4348bff-81f2-4e19-8098-bdac4cebc32f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.929657] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83c01fd2-efaf-4d53-9b92-1651318c1871 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.937310] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c904e31d-fdde-4947-be42-8970b035dad5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.942415] env[62460]: DEBUG nova.network.neutron [-] [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 534.967023] env[62460]: DEBUG nova.compute.provider_tree [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 534.970679] env[62460]: DEBUG nova.network.neutron [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 534.981152] env[62460]: DEBUG oslo_vmware.api [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': task-1313394, 'name': PowerOffVM_Task, 'duration_secs': 0.147891} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 534.981463] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 534.981702] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 534.982668] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e89add3-1f28-4aa7-9bbb-e58a1f1bfacf {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.995775] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 534.996130] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a98d4f2c-3734-4d13-8f8d-f63e5ac3b398 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.027018] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 535.027018] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Deleting contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 535.027018] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Deleting the datastore file [datastore1] 44a2ac83-3fff-4958-a1d3-b3884b42c211 {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 535.027018] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ba932ba8-5568-4e28-8828-0ac3d2c91ef3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.035851] env[62460]: DEBUG oslo_vmware.api [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Waiting for the task: (returnval){ [ 535.035851] env[62460]: value = "task-1313396" [ 535.035851] env[62460]: _type = "Task" [ 535.035851] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 535.045176] env[62460]: DEBUG oslo_vmware.api [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': task-1313396, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 535.345108] env[62460]: DEBUG nova.network.neutron [req-f7146609-282a-4a2a-b2a5-9d4843aec179 req-fc9e11cd-b0c6-495a-9ba7-124ea14a6efe service nova] [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 535.405173] env[62460]: DEBUG nova.network.neutron [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 535.444961] env[62460]: DEBUG nova.network.neutron [-] [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 535.472282] env[62460]: DEBUG nova.scheduler.client.report [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 535.544852] env[62460]: DEBUG nova.network.neutron [req-f7146609-282a-4a2a-b2a5-9d4843aec179 req-fc9e11cd-b0c6-495a-9ba7-124ea14a6efe service nova] [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 535.551231] env[62460]: DEBUG oslo_vmware.api [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': task-1313396, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.086079} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 535.551522] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 535.551715] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Deleted contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 535.551878] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 535.651115] env[62460]: DEBUG nova.compute.manager [req-9e41fed8-015c-40f6-85c1-d7de5a166565 req-1a41a09b-d332-40bf-83d2-ec569aa0a1bd service nova] [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] Received event network-changed-228ca897-3a3c-4e88-9838-451cdc94d456 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 535.651384] env[62460]: DEBUG nova.compute.manager [req-9e41fed8-015c-40f6-85c1-d7de5a166565 req-1a41a09b-d332-40bf-83d2-ec569aa0a1bd service nova] [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] Refreshing instance network info cache due to event network-changed-228ca897-3a3c-4e88-9838-451cdc94d456. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 535.652714] env[62460]: DEBUG oslo_concurrency.lockutils [req-9e41fed8-015c-40f6-85c1-d7de5a166565 req-1a41a09b-d332-40bf-83d2-ec569aa0a1bd service nova] Acquiring lock "refresh_cache-2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 535.665792] env[62460]: DEBUG nova.compute.manager [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 535.713645] env[62460]: DEBUG nova.virt.hardware [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 535.713891] env[62460]: DEBUG nova.virt.hardware [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 535.714063] env[62460]: DEBUG nova.virt.hardware [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 535.714253] env[62460]: DEBUG nova.virt.hardware [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 535.714403] env[62460]: DEBUG nova.virt.hardware [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 535.714607] env[62460]: DEBUG nova.virt.hardware [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 535.714788] env[62460]: DEBUG nova.virt.hardware [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 535.715093] env[62460]: DEBUG nova.virt.hardware [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 535.715209] env[62460]: DEBUG nova.virt.hardware [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 535.715377] env[62460]: DEBUG nova.virt.hardware [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 535.715559] env[62460]: DEBUG nova.virt.hardware [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 535.716464] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c725dea-e652-4536-b902-8ebc997b6e1d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.727490] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2fd407e-c625-4738-b7aa-7ea54260eb25 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.852553] env[62460]: DEBUG nova.network.neutron [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] Successfully created port: f6eb304f-e7db-40f7-a462-0db7083694d9 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 535.907857] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Releasing lock "refresh_cache-2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 535.908325] env[62460]: DEBUG nova.compute.manager [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 535.908523] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 535.908837] env[62460]: DEBUG oslo_concurrency.lockutils [req-9e41fed8-015c-40f6-85c1-d7de5a166565 req-1a41a09b-d332-40bf-83d2-ec569aa0a1bd service nova] Acquired lock "refresh_cache-2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 535.908992] env[62460]: DEBUG nova.network.neutron [req-9e41fed8-015c-40f6-85c1-d7de5a166565 req-1a41a09b-d332-40bf-83d2-ec569aa0a1bd service nova] [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] Refreshing network info cache for port 228ca897-3a3c-4e88-9838-451cdc94d456 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 535.910265] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cd4a8cc1-8a9f-465d-bfb7-615c5c8b324b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.921200] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1186094b-4d49-4840-8cc1-91eb3f1feba2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.945206] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f could not be found. [ 535.945206] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 535.945445] env[62460]: INFO nova.compute.manager [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 535.946018] env[62460]: DEBUG oslo.service.loopingcall [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 535.946018] env[62460]: DEBUG nova.compute.manager [-] [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 535.946018] env[62460]: DEBUG nova.network.neutron [-] [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 535.949290] env[62460]: INFO nova.compute.manager [-] [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] Took 1.10 seconds to deallocate network for instance. [ 535.953972] env[62460]: DEBUG nova.compute.claims [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 535.954255] env[62460]: DEBUG oslo_concurrency.lockutils [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 535.979188] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.345s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 535.979188] env[62460]: DEBUG nova.compute.manager [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 535.983590] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 4.628s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 536.046092] env[62460]: DEBUG oslo_concurrency.lockutils [req-f7146609-282a-4a2a-b2a5-9d4843aec179 req-fc9e11cd-b0c6-495a-9ba7-124ea14a6efe service nova] Releasing lock "refresh_cache-61d0d996-78ad-4b20-a7f4-b8a8f130d969" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 536.162492] env[62460]: DEBUG nova.network.neutron [-] [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 536.449785] env[62460]: DEBUG nova.network.neutron [req-9e41fed8-015c-40f6-85c1-d7de5a166565 req-1a41a09b-d332-40bf-83d2-ec569aa0a1bd service nova] [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 536.495509] env[62460]: DEBUG nova.compute.utils [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 536.501074] env[62460]: DEBUG nova.compute.manager [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 536.501377] env[62460]: DEBUG nova.network.neutron [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 536.567616] env[62460]: ERROR nova.compute.manager [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d8646801-5db8-41d8-a53c-f24e8f20d19a, please check neutron logs for more information. [ 536.567616] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 536.567616] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 536.567616] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 536.567616] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 536.567616] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 536.567616] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 536.567616] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 536.567616] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 536.567616] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 536.567616] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 536.567616] env[62460]: ERROR nova.compute.manager raise self.value [ 536.567616] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 536.567616] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 536.567616] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 536.567616] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 536.568139] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 536.568139] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 536.568139] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d8646801-5db8-41d8-a53c-f24e8f20d19a, please check neutron logs for more information. [ 536.568139] env[62460]: ERROR nova.compute.manager [ 536.568139] env[62460]: Traceback (most recent call last): [ 536.568139] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 536.568139] env[62460]: listener.cb(fileno) [ 536.568139] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 536.568139] env[62460]: result = function(*args, **kwargs) [ 536.568139] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 536.568139] env[62460]: return func(*args, **kwargs) [ 536.568139] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 536.568139] env[62460]: raise e [ 536.568139] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 536.568139] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 536.568139] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 536.568139] env[62460]: created_port_ids = self._update_ports_for_instance( [ 536.568139] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 536.568139] env[62460]: with excutils.save_and_reraise_exception(): [ 536.568139] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 536.568139] env[62460]: self.force_reraise() [ 536.568139] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 536.568139] env[62460]: raise self.value [ 536.568139] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 536.568139] env[62460]: updated_port = self._update_port( [ 536.568139] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 536.568139] env[62460]: _ensure_no_port_binding_failure(port) [ 536.568139] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 536.568139] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 536.570162] env[62460]: nova.exception.PortBindingFailed: Binding failed for port d8646801-5db8-41d8-a53c-f24e8f20d19a, please check neutron logs for more information. [ 536.570162] env[62460]: Removing descriptor: 16 [ 536.575989] env[62460]: ERROR nova.compute.manager [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] [instance: 7da5a822-6681-4dc6-b637-926ce425a042] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d8646801-5db8-41d8-a53c-f24e8f20d19a, please check neutron logs for more information. [ 536.575989] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] Traceback (most recent call last): [ 536.575989] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 536.575989] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] yield resources [ 536.575989] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 536.575989] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] self.driver.spawn(context, instance, image_meta, [ 536.575989] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 536.575989] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] self._vmops.spawn(context, instance, image_meta, injected_files, [ 536.575989] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 536.575989] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] vm_ref = self.build_virtual_machine(instance, [ 536.575989] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 536.576405] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] vif_infos = vmwarevif.get_vif_info(self._session, [ 536.576405] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 536.576405] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] for vif in network_info: [ 536.576405] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 536.576405] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] return self._sync_wrapper(fn, *args, **kwargs) [ 536.576405] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 536.576405] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] self.wait() [ 536.576405] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 536.576405] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] self[:] = self._gt.wait() [ 536.576405] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 536.576405] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] return self._exit_event.wait() [ 536.576405] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 536.576405] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] result = hub.switch() [ 536.576803] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 536.576803] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] return self.greenlet.switch() [ 536.576803] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 536.576803] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] result = function(*args, **kwargs) [ 536.576803] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 536.576803] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] return func(*args, **kwargs) [ 536.576803] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 536.576803] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] raise e [ 536.576803] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 536.576803] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] nwinfo = self.network_api.allocate_for_instance( [ 536.576803] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 536.576803] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] created_port_ids = self._update_ports_for_instance( [ 536.576803] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 536.577273] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] with excutils.save_and_reraise_exception(): [ 536.577273] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 536.577273] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] self.force_reraise() [ 536.577273] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 536.577273] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] raise self.value [ 536.577273] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 536.577273] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] updated_port = self._update_port( [ 536.577273] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 536.577273] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] _ensure_no_port_binding_failure(port) [ 536.577273] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 536.577273] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] raise exception.PortBindingFailed(port_id=port['id']) [ 536.577273] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] nova.exception.PortBindingFailed: Binding failed for port d8646801-5db8-41d8-a53c-f24e8f20d19a, please check neutron logs for more information. [ 536.577273] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] [ 536.577943] env[62460]: INFO nova.compute.manager [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] [instance: 7da5a822-6681-4dc6-b637-926ce425a042] Terminating instance [ 536.578978] env[62460]: DEBUG oslo_concurrency.lockutils [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] Acquiring lock "refresh_cache-7da5a822-6681-4dc6-b637-926ce425a042" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 536.579157] env[62460]: DEBUG oslo_concurrency.lockutils [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] Acquired lock "refresh_cache-7da5a822-6681-4dc6-b637-926ce425a042" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 536.579327] env[62460]: DEBUG nova.network.neutron [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] [instance: 7da5a822-6681-4dc6-b637-926ce425a042] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 536.605692] env[62460]: DEBUG nova.virt.hardware [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 536.605943] env[62460]: DEBUG nova.virt.hardware [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 536.606118] env[62460]: DEBUG nova.virt.hardware [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 536.606309] env[62460]: DEBUG nova.virt.hardware [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 536.606463] env[62460]: DEBUG nova.virt.hardware [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 536.606628] env[62460]: DEBUG nova.virt.hardware [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 536.606841] env[62460]: DEBUG nova.virt.hardware [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 536.607100] env[62460]: DEBUG nova.virt.hardware [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 536.607201] env[62460]: DEBUG nova.virt.hardware [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 536.607341] env[62460]: DEBUG nova.virt.hardware [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 536.607515] env[62460]: DEBUG nova.virt.hardware [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 536.608796] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5b90931-c134-46c8-aace-b38928a61006 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.619065] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9867f07c-d583-48a9-8edb-586302eff4a0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.632985] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Instance VIF info [] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 536.639311] env[62460]: DEBUG oslo.service.loopingcall [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 536.641017] env[62460]: DEBUG nova.policy [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cb788f92150a496f8a07966d151cddc5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '53051d5c63ce43098ee96c49d0855b3b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 536.642618] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 536.642784] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c2a880c1-b253-4462-bf1b-67227949acbd {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.659639] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 536.659639] env[62460]: value = "task-1313398" [ 536.659639] env[62460]: _type = "Task" [ 536.659639] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 536.663458] env[62460]: DEBUG nova.network.neutron [-] [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 536.668020] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313398, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 536.741273] env[62460]: DEBUG nova.network.neutron [req-9e41fed8-015c-40f6-85c1-d7de5a166565 req-1a41a09b-d332-40bf-83d2-ec569aa0a1bd service nova] [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 537.004210] env[62460]: DEBUG nova.compute.manager [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 537.040926] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 44a2ac83-3fff-4958-a1d3-b3884b42c211 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 537.041411] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 82afa576-f501-4a30-b8f7-e8b6334130cb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 537.041411] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 61d0d996-78ad-4b20-a7f4-b8a8f130d969 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 537.041411] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 7da5a822-6681-4dc6-b637-926ce425a042 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 537.043962] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 537.044059] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 073b92a2-8db9-4337-af2e-5e85c78d4b93 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 537.044299] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 537.044398] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance d16440e7-aa8a-4041-a4cd-d0ea84d8c027 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 537.117840] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Acquiring lock "ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 537.118100] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Lock "ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 537.130566] env[62460]: DEBUG nova.network.neutron [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] [instance: 7da5a822-6681-4dc6-b637-926ce425a042] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 537.169220] env[62460]: INFO nova.compute.manager [-] [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] Took 1.22 seconds to deallocate network for instance. [ 537.180019] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313398, 'name': CreateVM_Task, 'duration_secs': 0.348777} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 537.180019] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 537.180019] env[62460]: DEBUG nova.compute.claims [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 537.180019] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 537.181422] env[62460]: DEBUG oslo_vmware.service [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-882a4f5c-7235-4427-b0aa-bbbf75136f3f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.188149] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 537.188476] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 537.188987] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 537.189967] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-95235110-9705-4a5b-b0c0-1e8f92ed1a14 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.194312] env[62460]: DEBUG oslo_vmware.api [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Waiting for the task: (returnval){ [ 537.194312] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52028e8d-f02c-1003-bfdc-cc94b8112a57" [ 537.194312] env[62460]: _type = "Task" [ 537.194312] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 537.204349] env[62460]: DEBUG oslo_vmware.api [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52028e8d-f02c-1003-bfdc-cc94b8112a57, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 537.244748] env[62460]: DEBUG oslo_concurrency.lockutils [req-9e41fed8-015c-40f6-85c1-d7de5a166565 req-1a41a09b-d332-40bf-83d2-ec569aa0a1bd service nova] Releasing lock "refresh_cache-2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 537.405525] env[62460]: DEBUG nova.network.neutron [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] Successfully created port: 16f5912b-00e1-4f98-9484-4aabc70c360d {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 537.479009] env[62460]: DEBUG nova.network.neutron [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] [instance: 7da5a822-6681-4dc6-b637-926ce425a042] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 537.549109] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 72714a2e-a95d-495e-b163-f5168148fcdf has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 537.621820] env[62460]: DEBUG nova.compute.manager [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 537.713979] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 537.717017] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 537.717017] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 537.717017] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 537.717017] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 537.717017] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-196a0cd4-0c4f-4414-881a-faec03641a23 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.725819] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 537.725819] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 537.725819] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d69080cd-1fcb-431d-a590-363c8b7db467 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.734394] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-86a7d9af-722e-4692-a4ce-6c64aab72330 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.743498] env[62460]: DEBUG oslo_vmware.api [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Waiting for the task: (returnval){ [ 537.743498] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52934f25-cd62-841b-f10d-e4561ca996fa" [ 537.743498] env[62460]: _type = "Task" [ 537.743498] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 537.751869] env[62460]: DEBUG oslo_vmware.api [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52934f25-cd62-841b-f10d-e4561ca996fa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 537.828116] env[62460]: DEBUG nova.compute.manager [req-69d6a3e3-16e3-4094-8536-6ca05b239e7d req-7cc257ef-d16f-4b0b-943c-b693d4a1d4f6 service nova] [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] Received event network-vif-deleted-e30693e4-484a-4acc-ab76-e9da7cda9a03 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 537.828116] env[62460]: DEBUG nova.compute.manager [req-69d6a3e3-16e3-4094-8536-6ca05b239e7d req-7cc257ef-d16f-4b0b-943c-b693d4a1d4f6 service nova] [instance: 7da5a822-6681-4dc6-b637-926ce425a042] Received event network-changed-d8646801-5db8-41d8-a53c-f24e8f20d19a {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 537.828116] env[62460]: DEBUG nova.compute.manager [req-69d6a3e3-16e3-4094-8536-6ca05b239e7d req-7cc257ef-d16f-4b0b-943c-b693d4a1d4f6 service nova] [instance: 7da5a822-6681-4dc6-b637-926ce425a042] Refreshing instance network info cache due to event network-changed-d8646801-5db8-41d8-a53c-f24e8f20d19a. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 537.828116] env[62460]: DEBUG oslo_concurrency.lockutils [req-69d6a3e3-16e3-4094-8536-6ca05b239e7d req-7cc257ef-d16f-4b0b-943c-b693d4a1d4f6 service nova] Acquiring lock "refresh_cache-7da5a822-6681-4dc6-b637-926ce425a042" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 537.981651] env[62460]: DEBUG oslo_concurrency.lockutils [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] Releasing lock "refresh_cache-7da5a822-6681-4dc6-b637-926ce425a042" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 537.981651] env[62460]: DEBUG nova.compute.manager [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] [instance: 7da5a822-6681-4dc6-b637-926ce425a042] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 537.981651] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] [instance: 7da5a822-6681-4dc6-b637-926ce425a042] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 537.985244] env[62460]: DEBUG oslo_concurrency.lockutils [req-69d6a3e3-16e3-4094-8536-6ca05b239e7d req-7cc257ef-d16f-4b0b-943c-b693d4a1d4f6 service nova] Acquired lock "refresh_cache-7da5a822-6681-4dc6-b637-926ce425a042" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 537.985377] env[62460]: DEBUG nova.network.neutron [req-69d6a3e3-16e3-4094-8536-6ca05b239e7d req-7cc257ef-d16f-4b0b-943c-b693d4a1d4f6 service nova] [instance: 7da5a822-6681-4dc6-b637-926ce425a042] Refreshing network info cache for port d8646801-5db8-41d8-a53c-f24e8f20d19a {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 537.986517] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b6897ff2-e4a0-47b5-8b5d-e52c577db97b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.996147] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a771477d-8359-48a4-be68-d55aa90bd95e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.022746] env[62460]: DEBUG nova.compute.manager [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 538.025870] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] [instance: 7da5a822-6681-4dc6-b637-926ce425a042] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7da5a822-6681-4dc6-b637-926ce425a042 could not be found. [ 538.025959] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] [instance: 7da5a822-6681-4dc6-b637-926ce425a042] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 538.026085] env[62460]: INFO nova.compute.manager [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] [instance: 7da5a822-6681-4dc6-b637-926ce425a042] Took 0.04 seconds to destroy the instance on the hypervisor. [ 538.026340] env[62460]: DEBUG oslo.service.loopingcall [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 538.027305] env[62460]: DEBUG nova.compute.manager [-] [instance: 7da5a822-6681-4dc6-b637-926ce425a042] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 538.027374] env[62460]: DEBUG nova.network.neutron [-] [instance: 7da5a822-6681-4dc6-b637-926ce425a042] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 538.058572] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 538.060512] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=62460) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 538.060512] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2048MB phys_disk=200GB used_disk=8GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=62460) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 538.071175] env[62460]: DEBUG nova.virt.hardware [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T09:00:04Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='1401376678',id=23,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-409029658',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 538.071175] env[62460]: DEBUG nova.virt.hardware [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 538.071175] env[62460]: DEBUG nova.virt.hardware [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 538.071467] env[62460]: DEBUG nova.virt.hardware [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 538.071467] env[62460]: DEBUG nova.virt.hardware [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 538.071467] env[62460]: DEBUG nova.virt.hardware [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 538.071651] env[62460]: DEBUG nova.virt.hardware [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 538.071793] env[62460]: DEBUG nova.virt.hardware [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 538.072493] env[62460]: DEBUG nova.virt.hardware [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 538.073293] env[62460]: DEBUG nova.virt.hardware [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 538.073843] env[62460]: DEBUG nova.virt.hardware [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 538.075773] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61f818b9-079b-4bb3-9b63-146d2842e2a0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.086777] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d83a621d-ca10-4902-b36f-ba9aaf5d6ffc {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.092765] env[62460]: DEBUG nova.network.neutron [-] [instance: 7da5a822-6681-4dc6-b637-926ce425a042] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 538.150016] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 538.188723] env[62460]: ERROR nova.compute.manager [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6d1179e6-551c-4986-b849-eceeefc23045, please check neutron logs for more information. [ 538.188723] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 538.188723] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 538.188723] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 538.188723] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 538.188723] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 538.188723] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 538.188723] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 538.188723] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 538.188723] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 538.188723] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 538.188723] env[62460]: ERROR nova.compute.manager raise self.value [ 538.188723] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 538.188723] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 538.188723] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 538.188723] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 538.189793] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 538.189793] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 538.189793] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6d1179e6-551c-4986-b849-eceeefc23045, please check neutron logs for more information. [ 538.189793] env[62460]: ERROR nova.compute.manager [ 538.189793] env[62460]: Traceback (most recent call last): [ 538.189793] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 538.189793] env[62460]: listener.cb(fileno) [ 538.189793] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 538.189793] env[62460]: result = function(*args, **kwargs) [ 538.189793] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 538.189793] env[62460]: return func(*args, **kwargs) [ 538.189793] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 538.189793] env[62460]: raise e [ 538.189793] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 538.189793] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 538.189793] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 538.189793] env[62460]: created_port_ids = self._update_ports_for_instance( [ 538.189793] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 538.189793] env[62460]: with excutils.save_and_reraise_exception(): [ 538.189793] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 538.189793] env[62460]: self.force_reraise() [ 538.189793] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 538.189793] env[62460]: raise self.value [ 538.189793] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 538.189793] env[62460]: updated_port = self._update_port( [ 538.189793] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 538.189793] env[62460]: _ensure_no_port_binding_failure(port) [ 538.189793] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 538.189793] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 538.191239] env[62460]: nova.exception.PortBindingFailed: Binding failed for port 6d1179e6-551c-4986-b849-eceeefc23045, please check neutron logs for more information. [ 538.191239] env[62460]: Removing descriptor: 15 [ 538.192174] env[62460]: ERROR nova.compute.manager [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6d1179e6-551c-4986-b849-eceeefc23045, please check neutron logs for more information. [ 538.192174] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] Traceback (most recent call last): [ 538.192174] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 538.192174] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] yield resources [ 538.192174] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 538.192174] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] self.driver.spawn(context, instance, image_meta, [ 538.192174] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 538.192174] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] self._vmops.spawn(context, instance, image_meta, injected_files, [ 538.192174] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 538.192174] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] vm_ref = self.build_virtual_machine(instance, [ 538.192174] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 538.192711] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] vif_infos = vmwarevif.get_vif_info(self._session, [ 538.192711] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 538.192711] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] for vif in network_info: [ 538.192711] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 538.192711] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] return self._sync_wrapper(fn, *args, **kwargs) [ 538.192711] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 538.192711] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] self.wait() [ 538.192711] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 538.192711] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] self[:] = self._gt.wait() [ 538.192711] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 538.192711] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] return self._exit_event.wait() [ 538.192711] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 538.192711] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] result = hub.switch() [ 538.194284] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 538.194284] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] return self.greenlet.switch() [ 538.194284] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 538.194284] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] result = function(*args, **kwargs) [ 538.194284] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 538.194284] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] return func(*args, **kwargs) [ 538.194284] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 538.194284] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] raise e [ 538.194284] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 538.194284] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] nwinfo = self.network_api.allocate_for_instance( [ 538.194284] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 538.194284] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] created_port_ids = self._update_ports_for_instance( [ 538.194284] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 538.195129] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] with excutils.save_and_reraise_exception(): [ 538.195129] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 538.195129] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] self.force_reraise() [ 538.195129] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 538.195129] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] raise self.value [ 538.195129] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 538.195129] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] updated_port = self._update_port( [ 538.195129] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 538.195129] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] _ensure_no_port_binding_failure(port) [ 538.195129] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 538.195129] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] raise exception.PortBindingFailed(port_id=port['id']) [ 538.195129] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] nova.exception.PortBindingFailed: Binding failed for port 6d1179e6-551c-4986-b849-eceeefc23045, please check neutron logs for more information. [ 538.195129] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] [ 538.196363] env[62460]: INFO nova.compute.manager [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] Terminating instance [ 538.198441] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Acquiring lock "refresh_cache-073b92a2-8db9-4337-af2e-5e85c78d4b93" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 538.198441] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Acquired lock "refresh_cache-073b92a2-8db9-4337-af2e-5e85c78d4b93" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 538.198441] env[62460]: DEBUG nova.network.neutron [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 538.257077] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Preparing fetch location {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 538.257643] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Creating directory with path [datastore2] vmware_temp/7f279b10-e92a-494e-92db-635315503694/eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 538.257643] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-283b057c-681d-40cf-9f12-9727940c25df {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.270209] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25385721-ff26-48c4-ad42-85abd162a845 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.277156] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3066bda-2f5c-4a44-bb93-00489c42733e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.313028] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3556a7ae-59c1-4233-9521-ba16b5c5fd9a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.315762] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Created directory with path [datastore2] vmware_temp/7f279b10-e92a-494e-92db-635315503694/eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 538.315987] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Fetch image to [datastore2] vmware_temp/7f279b10-e92a-494e-92db-635315503694/eb3fb456-2dfd-4d95-978f-f17699d7735d/tmp-sparse.vmdk {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 538.316183] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Downloading image file data eb3fb456-2dfd-4d95-978f-f17699d7735d to [datastore2] vmware_temp/7f279b10-e92a-494e-92db-635315503694/eb3fb456-2dfd-4d95-978f-f17699d7735d/tmp-sparse.vmdk on the data store datastore2 {{(pid=62460) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 538.317153] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27b80a32-e3b8-4cf8-9ecc-88497c3616cb {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.328306] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b0ecc0b-f336-4f04-9ec5-a992b4788ef7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.331014] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ab86f2c-24aa-4feb-9661-a430fdd9512f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.354622] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b14ee8f5-bb0e-422c-ae0b-1c66a3a8d215 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.359795] env[62460]: DEBUG nova.compute.provider_tree [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 538.394357] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb572c09-4a15-404e-94e9-70118b88ac2a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.402456] env[62460]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-314c9ea2-4c2f-49dd-b730-8aa4668ccddb {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.494945] env[62460]: DEBUG nova.virt.vmwareapi.images [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Downloading image file data eb3fb456-2dfd-4d95-978f-f17699d7735d to the data store datastore2 {{(pid=62460) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 538.518113] env[62460]: DEBUG nova.network.neutron [req-69d6a3e3-16e3-4094-8536-6ca05b239e7d req-7cc257ef-d16f-4b0b-943c-b693d4a1d4f6 service nova] [instance: 7da5a822-6681-4dc6-b637-926ce425a042] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 538.565159] env[62460]: DEBUG oslo_vmware.rw_handles [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/7f279b10-e92a-494e-92db-635315503694/eb3fb456-2dfd-4d95-978f-f17699d7735d/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=62460) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 538.625749] env[62460]: DEBUG nova.network.neutron [-] [instance: 7da5a822-6681-4dc6-b637-926ce425a042] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 538.746569] env[62460]: DEBUG nova.network.neutron [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 538.781358] env[62460]: DEBUG nova.network.neutron [req-69d6a3e3-16e3-4094-8536-6ca05b239e7d req-7cc257ef-d16f-4b0b-943c-b693d4a1d4f6 service nova] [instance: 7da5a822-6681-4dc6-b637-926ce425a042] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 538.863026] env[62460]: DEBUG nova.scheduler.client.report [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 538.921140] env[62460]: DEBUG nova.network.neutron [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 539.133857] env[62460]: INFO nova.compute.manager [-] [instance: 7da5a822-6681-4dc6-b637-926ce425a042] Took 1.10 seconds to deallocate network for instance. [ 539.135916] env[62460]: DEBUG nova.compute.claims [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] [instance: 7da5a822-6681-4dc6-b637-926ce425a042] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 539.136128] env[62460]: DEBUG oslo_concurrency.lockutils [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.159802] env[62460]: DEBUG nova.compute.manager [req-97cdaac7-fd6b-4960-a933-c0e7f872a1d4 req-b14d9a0f-fc57-45ca-b37e-164840514880 service nova] [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] Received event network-vif-deleted-228ca897-3a3c-4e88-9838-451cdc94d456 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 539.288202] env[62460]: DEBUG oslo_concurrency.lockutils [req-69d6a3e3-16e3-4094-8536-6ca05b239e7d req-7cc257ef-d16f-4b0b-943c-b693d4a1d4f6 service nova] Releasing lock "refresh_cache-7da5a822-6681-4dc6-b637-926ce425a042" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 539.288776] env[62460]: DEBUG nova.compute.manager [req-69d6a3e3-16e3-4094-8536-6ca05b239e7d req-7cc257ef-d16f-4b0b-943c-b693d4a1d4f6 service nova] [instance: 7da5a822-6681-4dc6-b637-926ce425a042] Received event network-vif-deleted-d8646801-5db8-41d8-a53c-f24e8f20d19a {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 539.316506] env[62460]: DEBUG oslo_vmware.rw_handles [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Completed reading data from the image iterator. {{(pid=62460) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 539.316506] env[62460]: DEBUG oslo_vmware.rw_handles [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Closing write handle for https://esx7c2n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/7f279b10-e92a-494e-92db-635315503694/eb3fb456-2dfd-4d95-978f-f17699d7735d/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=62460) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 539.371988] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62460) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 539.372259] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.389s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 539.372549] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.972s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.374270] env[62460]: INFO nova.compute.claims [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 539.389858] env[62460]: DEBUG nova.virt.vmwareapi.images [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Downloaded image file data eb3fb456-2dfd-4d95-978f-f17699d7735d to vmware_temp/7f279b10-e92a-494e-92db-635315503694/eb3fb456-2dfd-4d95-978f-f17699d7735d/tmp-sparse.vmdk on the data store datastore2 {{(pid=62460) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 539.391086] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Caching image {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 539.391341] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Copying Virtual Disk [datastore2] vmware_temp/7f279b10-e92a-494e-92db-635315503694/eb3fb456-2dfd-4d95-978f-f17699d7735d/tmp-sparse.vmdk to [datastore2] vmware_temp/7f279b10-e92a-494e-92db-635315503694/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 539.391629] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-076f7ff3-a79e-46c0-ac57-3637bfd0a315 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.402817] env[62460]: DEBUG oslo_vmware.api [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Waiting for the task: (returnval){ [ 539.402817] env[62460]: value = "task-1313400" [ 539.402817] env[62460]: _type = "Task" [ 539.402817] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 539.416697] env[62460]: DEBUG oslo_vmware.api [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': task-1313400, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 539.424582] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Releasing lock "refresh_cache-073b92a2-8db9-4337-af2e-5e85c78d4b93" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 539.425095] env[62460]: DEBUG nova.compute.manager [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 539.425394] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 539.426178] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-74f2e3e0-da49-4f12-93f5-6f146734b65e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.434861] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ff6df95-febf-4a71-a7bb-1ae56e15ae7b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.460083] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 073b92a2-8db9-4337-af2e-5e85c78d4b93 could not be found. [ 539.460404] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 539.460745] env[62460]: INFO nova.compute.manager [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] Took 0.04 seconds to destroy the instance on the hypervisor. [ 539.461198] env[62460]: DEBUG oslo.service.loopingcall [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 539.462350] env[62460]: DEBUG nova.compute.manager [-] [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 539.462981] env[62460]: DEBUG nova.network.neutron [-] [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 539.495018] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] Acquiring lock "d52580bc-7c11-47f0-a6c4-12941c48d75c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.495018] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] Lock "d52580bc-7c11-47f0-a6c4-12941c48d75c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.515107] env[62460]: DEBUG nova.network.neutron [-] [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 539.913642] env[62460]: DEBUG oslo_vmware.api [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': task-1313400, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 539.995099] env[62460]: DEBUG nova.compute.manager [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 540.016573] env[62460]: ERROR nova.compute.manager [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 16f5912b-00e1-4f98-9484-4aabc70c360d, please check neutron logs for more information. [ 540.016573] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 540.016573] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 540.016573] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 540.016573] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 540.016573] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 540.016573] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 540.016573] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 540.016573] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 540.016573] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 540.016573] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 540.016573] env[62460]: ERROR nova.compute.manager raise self.value [ 540.016573] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 540.016573] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 540.016573] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 540.016573] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 540.017022] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 540.017022] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 540.017022] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 16f5912b-00e1-4f98-9484-4aabc70c360d, please check neutron logs for more information. [ 540.017022] env[62460]: ERROR nova.compute.manager [ 540.017022] env[62460]: Traceback (most recent call last): [ 540.017022] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 540.017022] env[62460]: listener.cb(fileno) [ 540.017022] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 540.017022] env[62460]: result = function(*args, **kwargs) [ 540.017022] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 540.017022] env[62460]: return func(*args, **kwargs) [ 540.017022] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 540.017022] env[62460]: raise e [ 540.017022] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 540.017022] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 540.017022] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 540.017022] env[62460]: created_port_ids = self._update_ports_for_instance( [ 540.017022] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 540.017022] env[62460]: with excutils.save_and_reraise_exception(): [ 540.017022] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 540.017022] env[62460]: self.force_reraise() [ 540.017022] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 540.017022] env[62460]: raise self.value [ 540.017022] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 540.017022] env[62460]: updated_port = self._update_port( [ 540.017022] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 540.017022] env[62460]: _ensure_no_port_binding_failure(port) [ 540.017022] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 540.017022] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 540.017753] env[62460]: nova.exception.PortBindingFailed: Binding failed for port 16f5912b-00e1-4f98-9484-4aabc70c360d, please check neutron logs for more information. [ 540.017753] env[62460]: Removing descriptor: 18 [ 540.017753] env[62460]: ERROR nova.compute.manager [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 16f5912b-00e1-4f98-9484-4aabc70c360d, please check neutron logs for more information. [ 540.017753] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] Traceback (most recent call last): [ 540.017753] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 540.017753] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] yield resources [ 540.017753] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 540.017753] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] self.driver.spawn(context, instance, image_meta, [ 540.017753] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 540.017753] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] self._vmops.spawn(context, instance, image_meta, injected_files, [ 540.017753] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 540.017753] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] vm_ref = self.build_virtual_machine(instance, [ 540.018079] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 540.018079] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] vif_infos = vmwarevif.get_vif_info(self._session, [ 540.018079] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 540.018079] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] for vif in network_info: [ 540.018079] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 540.018079] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] return self._sync_wrapper(fn, *args, **kwargs) [ 540.018079] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 540.018079] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] self.wait() [ 540.018079] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 540.018079] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] self[:] = self._gt.wait() [ 540.018079] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 540.018079] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] return self._exit_event.wait() [ 540.018079] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 540.018426] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] result = hub.switch() [ 540.018426] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 540.018426] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] return self.greenlet.switch() [ 540.018426] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 540.018426] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] result = function(*args, **kwargs) [ 540.018426] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 540.018426] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] return func(*args, **kwargs) [ 540.018426] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 540.018426] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] raise e [ 540.018426] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 540.018426] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] nwinfo = self.network_api.allocate_for_instance( [ 540.018426] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 540.018426] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] created_port_ids = self._update_ports_for_instance( [ 540.018745] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 540.018745] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] with excutils.save_and_reraise_exception(): [ 540.018745] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 540.018745] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] self.force_reraise() [ 540.018745] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 540.018745] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] raise self.value [ 540.018745] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 540.018745] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] updated_port = self._update_port( [ 540.018745] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 540.018745] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] _ensure_no_port_binding_failure(port) [ 540.018745] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 540.018745] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] raise exception.PortBindingFailed(port_id=port['id']) [ 540.019085] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] nova.exception.PortBindingFailed: Binding failed for port 16f5912b-00e1-4f98-9484-4aabc70c360d, please check neutron logs for more information. [ 540.019085] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] [ 540.019085] env[62460]: INFO nova.compute.manager [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] Terminating instance [ 540.020207] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] Acquiring lock "refresh_cache-d16440e7-aa8a-4041-a4cd-d0ea84d8c027" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 540.020281] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] Acquired lock "refresh_cache-d16440e7-aa8a-4041-a4cd-d0ea84d8c027" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 540.020412] env[62460]: DEBUG nova.network.neutron [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 540.021811] env[62460]: DEBUG nova.network.neutron [-] [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 540.324153] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Acquiring lock "00544d2a-1a15-4347-abe4-3641b5bf0cfe" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.324459] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Lock "00544d2a-1a15-4347-abe4-3641b5bf0cfe" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 540.420053] env[62460]: DEBUG oslo_vmware.api [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': task-1313400, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.679014} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 540.420571] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Copied Virtual Disk [datastore2] vmware_temp/7f279b10-e92a-494e-92db-635315503694/eb3fb456-2dfd-4d95-978f-f17699d7735d/tmp-sparse.vmdk to [datastore2] vmware_temp/7f279b10-e92a-494e-92db-635315503694/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 540.420928] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Deleting the datastore file [datastore2] vmware_temp/7f279b10-e92a-494e-92db-635315503694/eb3fb456-2dfd-4d95-978f-f17699d7735d/tmp-sparse.vmdk {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 540.421497] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7468b811-aaa2-4f74-afa4-731243ba6d60 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.430489] env[62460]: DEBUG oslo_vmware.api [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Waiting for the task: (returnval){ [ 540.430489] env[62460]: value = "task-1313401" [ 540.430489] env[62460]: _type = "Task" [ 540.430489] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 540.439982] env[62460]: DEBUG oslo_vmware.api [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': task-1313401, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 540.516670] env[62460]: DEBUG nova.compute.manager [req-082bb4f5-9d0d-40cf-b48e-54f2db013416 req-7a251895-e105-40d4-994c-c68461b669b4 service nova] [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] Received event network-changed-6d1179e6-551c-4986-b849-eceeefc23045 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 540.516868] env[62460]: DEBUG nova.compute.manager [req-082bb4f5-9d0d-40cf-b48e-54f2db013416 req-7a251895-e105-40d4-994c-c68461b669b4 service nova] [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] Refreshing instance network info cache due to event network-changed-6d1179e6-551c-4986-b849-eceeefc23045. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 540.517012] env[62460]: DEBUG oslo_concurrency.lockutils [req-082bb4f5-9d0d-40cf-b48e-54f2db013416 req-7a251895-e105-40d4-994c-c68461b669b4 service nova] Acquiring lock "refresh_cache-073b92a2-8db9-4337-af2e-5e85c78d4b93" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 540.517118] env[62460]: DEBUG oslo_concurrency.lockutils [req-082bb4f5-9d0d-40cf-b48e-54f2db013416 req-7a251895-e105-40d4-994c-c68461b669b4 service nova] Acquired lock "refresh_cache-073b92a2-8db9-4337-af2e-5e85c78d4b93" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 540.517249] env[62460]: DEBUG nova.network.neutron [req-082bb4f5-9d0d-40cf-b48e-54f2db013416 req-7a251895-e105-40d4-994c-c68461b669b4 service nova] [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] Refreshing network info cache for port 6d1179e6-551c-4986-b849-eceeefc23045 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 540.528400] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.532892] env[62460]: INFO nova.compute.manager [-] [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] Took 1.07 seconds to deallocate network for instance. [ 540.535403] env[62460]: DEBUG nova.compute.claims [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 540.535606] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.582189] env[62460]: DEBUG nova.network.neutron [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 540.589991] env[62460]: ERROR nova.compute.manager [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f6eb304f-e7db-40f7-a462-0db7083694d9, please check neutron logs for more information. [ 540.589991] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 540.589991] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 540.589991] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 540.589991] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 540.589991] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 540.589991] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 540.589991] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 540.589991] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 540.589991] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 540.589991] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 540.589991] env[62460]: ERROR nova.compute.manager raise self.value [ 540.589991] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 540.589991] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 540.589991] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 540.589991] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 540.590584] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 540.590584] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 540.590584] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f6eb304f-e7db-40f7-a462-0db7083694d9, please check neutron logs for more information. [ 540.590584] env[62460]: ERROR nova.compute.manager [ 540.590584] env[62460]: Traceback (most recent call last): [ 540.590584] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 540.590584] env[62460]: listener.cb(fileno) [ 540.590584] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 540.590584] env[62460]: result = function(*args, **kwargs) [ 540.590584] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 540.590584] env[62460]: return func(*args, **kwargs) [ 540.590584] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 540.590584] env[62460]: raise e [ 540.590584] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 540.590584] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 540.590584] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 540.590584] env[62460]: created_port_ids = self._update_ports_for_instance( [ 540.590584] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 540.590584] env[62460]: with excutils.save_and_reraise_exception(): [ 540.590584] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 540.590584] env[62460]: self.force_reraise() [ 540.590584] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 540.590584] env[62460]: raise self.value [ 540.590584] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 540.590584] env[62460]: updated_port = self._update_port( [ 540.590584] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 540.590584] env[62460]: _ensure_no_port_binding_failure(port) [ 540.590584] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 540.590584] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 540.591369] env[62460]: nova.exception.PortBindingFailed: Binding failed for port f6eb304f-e7db-40f7-a462-0db7083694d9, please check neutron logs for more information. [ 540.591369] env[62460]: Removing descriptor: 19 [ 540.591369] env[62460]: ERROR nova.compute.manager [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f6eb304f-e7db-40f7-a462-0db7083694d9, please check neutron logs for more information. [ 540.591369] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] Traceback (most recent call last): [ 540.591369] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 540.591369] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] yield resources [ 540.591369] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 540.591369] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] self.driver.spawn(context, instance, image_meta, [ 540.591369] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 540.591369] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] self._vmops.spawn(context, instance, image_meta, injected_files, [ 540.591369] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 540.591369] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] vm_ref = self.build_virtual_machine(instance, [ 540.591691] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 540.591691] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] vif_infos = vmwarevif.get_vif_info(self._session, [ 540.591691] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 540.591691] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] for vif in network_info: [ 540.591691] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 540.591691] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] return self._sync_wrapper(fn, *args, **kwargs) [ 540.591691] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 540.591691] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] self.wait() [ 540.591691] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 540.591691] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] self[:] = self._gt.wait() [ 540.591691] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 540.591691] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] return self._exit_event.wait() [ 540.591691] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 540.592061] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] result = hub.switch() [ 540.592061] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 540.592061] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] return self.greenlet.switch() [ 540.592061] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 540.592061] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] result = function(*args, **kwargs) [ 540.592061] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 540.592061] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] return func(*args, **kwargs) [ 540.592061] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 540.592061] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] raise e [ 540.592061] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 540.592061] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] nwinfo = self.network_api.allocate_for_instance( [ 540.592061] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 540.592061] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] created_port_ids = self._update_ports_for_instance( [ 540.592480] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 540.592480] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] with excutils.save_and_reraise_exception(): [ 540.592480] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 540.592480] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] self.force_reraise() [ 540.592480] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 540.592480] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] raise self.value [ 540.592480] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 540.592480] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] updated_port = self._update_port( [ 540.592480] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 540.592480] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] _ensure_no_port_binding_failure(port) [ 540.592480] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 540.592480] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] raise exception.PortBindingFailed(port_id=port['id']) [ 540.592778] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] nova.exception.PortBindingFailed: Binding failed for port f6eb304f-e7db-40f7-a462-0db7083694d9, please check neutron logs for more information. [ 540.592778] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] [ 540.592778] env[62460]: INFO nova.compute.manager [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] Terminating instance [ 540.594834] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Acquiring lock "refresh_cache-48e50ab0-9d4e-4c5d-96d5-55399b3a8e19" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 540.594996] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Acquired lock "refresh_cache-48e50ab0-9d4e-4c5d-96d5-55399b3a8e19" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 540.595181] env[62460]: DEBUG nova.network.neutron [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 540.621891] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-854b5e62-76fb-45ed-8925-2317730a5cfd {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.633451] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c4f9f9a-6e9c-4374-91f5-135747e4137e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.669872] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4231411-36ee-4f33-8f8d-c42caac9e426 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.678301] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-291c9dc1-87b5-4786-b954-131f1d2ae912 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.692072] env[62460]: DEBUG nova.compute.provider_tree [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 540.862175] env[62460]: DEBUG nova.network.neutron [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 540.948709] env[62460]: DEBUG oslo_vmware.api [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': task-1313401, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.023207} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 540.949257] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 540.949563] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Moving file from [datastore2] vmware_temp/7f279b10-e92a-494e-92db-635315503694/eb3fb456-2dfd-4d95-978f-f17699d7735d to [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d. {{(pid=62460) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 540.950401] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-96e75e75-0b93-4fa3-a064-16132ccccbe6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.962188] env[62460]: DEBUG oslo_vmware.api [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Waiting for the task: (returnval){ [ 540.962188] env[62460]: value = "task-1313403" [ 540.962188] env[62460]: _type = "Task" [ 540.962188] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 540.972755] env[62460]: DEBUG oslo_vmware.api [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': task-1313403, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 541.055727] env[62460]: DEBUG nova.network.neutron [req-082bb4f5-9d0d-40cf-b48e-54f2db013416 req-7a251895-e105-40d4-994c-c68461b669b4 service nova] [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 541.165659] env[62460]: DEBUG nova.network.neutron [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 541.198038] env[62460]: DEBUG nova.scheduler.client.report [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 541.350424] env[62460]: DEBUG nova.network.neutron [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 541.355894] env[62460]: DEBUG nova.network.neutron [req-082bb4f5-9d0d-40cf-b48e-54f2db013416 req-7a251895-e105-40d4-994c-c68461b669b4 service nova] [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 541.364395] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] Releasing lock "refresh_cache-d16440e7-aa8a-4041-a4cd-d0ea84d8c027" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 541.364722] env[62460]: DEBUG nova.compute.manager [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 541.365936] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 541.365936] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9710da24-4d43-4c64-803a-b9c6870232f5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.378920] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d6fa8f3-20ea-435b-b19f-b0bc61fbe175 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.413971] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d16440e7-aa8a-4041-a4cd-d0ea84d8c027 could not be found. [ 541.414251] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 541.414439] env[62460]: INFO nova.compute.manager [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] Took 0.05 seconds to destroy the instance on the hypervisor. [ 541.414679] env[62460]: DEBUG oslo.service.loopingcall [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 541.414972] env[62460]: DEBUG nova.compute.manager [-] [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 541.414972] env[62460]: DEBUG nova.network.neutron [-] [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 541.474565] env[62460]: DEBUG oslo_vmware.api [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': task-1313403, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.142953} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 541.474837] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] File moved {{(pid=62460) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 541.475044] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Cleaning up location [datastore2] vmware_temp/7f279b10-e92a-494e-92db-635315503694 {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 541.475298] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Deleting the datastore file [datastore2] vmware_temp/7f279b10-e92a-494e-92db-635315503694 {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 541.475457] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3fd520e7-d510-4aa4-8398-1089df1836fc {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.483483] env[62460]: DEBUG oslo_vmware.api [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Waiting for the task: (returnval){ [ 541.483483] env[62460]: value = "task-1313404" [ 541.483483] env[62460]: _type = "Task" [ 541.483483] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 541.495382] env[62460]: DEBUG oslo_vmware.api [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': task-1313404, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 541.624957] env[62460]: DEBUG nova.network.neutron [-] [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 541.710209] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.338s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 541.711788] env[62460]: DEBUG nova.compute.manager [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 541.715374] env[62460]: DEBUG oslo_concurrency.lockutils [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 8.565s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 541.801946] env[62460]: DEBUG nova.compute.manager [req-5811884b-3c39-4e59-8d84-238eb67cc25d req-34e5a7c9-7c76-475e-9ad3-e9ebbf55b129 service nova] [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] Received event network-changed-f6eb304f-e7db-40f7-a462-0db7083694d9 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 541.802324] env[62460]: DEBUG nova.compute.manager [req-5811884b-3c39-4e59-8d84-238eb67cc25d req-34e5a7c9-7c76-475e-9ad3-e9ebbf55b129 service nova] [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] Refreshing instance network info cache due to event network-changed-f6eb304f-e7db-40f7-a462-0db7083694d9. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 541.802625] env[62460]: DEBUG oslo_concurrency.lockutils [req-5811884b-3c39-4e59-8d84-238eb67cc25d req-34e5a7c9-7c76-475e-9ad3-e9ebbf55b129 service nova] Acquiring lock "refresh_cache-48e50ab0-9d4e-4c5d-96d5-55399b3a8e19" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 541.853146] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Releasing lock "refresh_cache-48e50ab0-9d4e-4c5d-96d5-55399b3a8e19" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 541.853569] env[62460]: DEBUG nova.compute.manager [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 541.853769] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 541.853980] env[62460]: DEBUG oslo_concurrency.lockutils [req-5811884b-3c39-4e59-8d84-238eb67cc25d req-34e5a7c9-7c76-475e-9ad3-e9ebbf55b129 service nova] Acquired lock "refresh_cache-48e50ab0-9d4e-4c5d-96d5-55399b3a8e19" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 541.854252] env[62460]: DEBUG nova.network.neutron [req-5811884b-3c39-4e59-8d84-238eb67cc25d req-34e5a7c9-7c76-475e-9ad3-e9ebbf55b129 service nova] [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] Refreshing network info cache for port f6eb304f-e7db-40f7-a462-0db7083694d9 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 541.855249] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1eb2fad4-db28-46db-8ccd-4dfc34603f5b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.865898] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-719b623c-b4e6-4503-95c6-c80698676b87 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.879240] env[62460]: DEBUG oslo_concurrency.lockutils [req-082bb4f5-9d0d-40cf-b48e-54f2db013416 req-7a251895-e105-40d4-994c-c68461b669b4 service nova] Releasing lock "refresh_cache-073b92a2-8db9-4337-af2e-5e85c78d4b93" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 541.879658] env[62460]: DEBUG nova.compute.manager [req-082bb4f5-9d0d-40cf-b48e-54f2db013416 req-7a251895-e105-40d4-994c-c68461b669b4 service nova] [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] Received event network-vif-deleted-6d1179e6-551c-4986-b849-eceeefc23045 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 541.879955] env[62460]: DEBUG nova.compute.manager [req-082bb4f5-9d0d-40cf-b48e-54f2db013416 req-7a251895-e105-40d4-994c-c68461b669b4 service nova] [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] Received event network-changed-16f5912b-00e1-4f98-9484-4aabc70c360d {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 541.880263] env[62460]: DEBUG nova.compute.manager [req-082bb4f5-9d0d-40cf-b48e-54f2db013416 req-7a251895-e105-40d4-994c-c68461b669b4 service nova] [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] Refreshing instance network info cache due to event network-changed-16f5912b-00e1-4f98-9484-4aabc70c360d. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 541.880612] env[62460]: DEBUG oslo_concurrency.lockutils [req-082bb4f5-9d0d-40cf-b48e-54f2db013416 req-7a251895-e105-40d4-994c-c68461b669b4 service nova] Acquiring lock "refresh_cache-d16440e7-aa8a-4041-a4cd-d0ea84d8c027" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 541.880866] env[62460]: DEBUG oslo_concurrency.lockutils [req-082bb4f5-9d0d-40cf-b48e-54f2db013416 req-7a251895-e105-40d4-994c-c68461b669b4 service nova] Acquired lock "refresh_cache-d16440e7-aa8a-4041-a4cd-d0ea84d8c027" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 541.881150] env[62460]: DEBUG nova.network.neutron [req-082bb4f5-9d0d-40cf-b48e-54f2db013416 req-7a251895-e105-40d4-994c-c68461b669b4 service nova] [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] Refreshing network info cache for port 16f5912b-00e1-4f98-9484-4aabc70c360d {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 541.898622] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19 could not be found. [ 541.898622] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 541.898622] env[62460]: INFO nova.compute.manager [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] Took 0.04 seconds to destroy the instance on the hypervisor. [ 541.898781] env[62460]: DEBUG oslo.service.loopingcall [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 541.898882] env[62460]: DEBUG nova.compute.manager [-] [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 541.898985] env[62460]: DEBUG nova.network.neutron [-] [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 541.972385] env[62460]: DEBUG nova.network.neutron [-] [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 541.998307] env[62460]: DEBUG oslo_vmware.api [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': task-1313404, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.029871} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 541.998713] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 541.999475] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b0ffe8d8-7451-44cd-b643-e5de8bc143ce {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.005881] env[62460]: DEBUG oslo_vmware.api [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Waiting for the task: (returnval){ [ 542.005881] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52f6bbd2-71f3-0d90-33d3-a7fcec45bf84" [ 542.005881] env[62460]: _type = "Task" [ 542.005881] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 542.020542] env[62460]: DEBUG oslo_vmware.api [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52f6bbd2-71f3-0d90-33d3-a7fcec45bf84, 'name': SearchDatastore_Task, 'duration_secs': 0.009276} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 542.020829] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 542.021107] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] 44a2ac83-3fff-4958-a1d3-b3884b42c211/44a2ac83-3fff-4958-a1d3-b3884b42c211.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 542.021514] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a9dc28f3-bd3d-4352-8fa1-6b45f8252703 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.030078] env[62460]: DEBUG oslo_vmware.api [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Waiting for the task: (returnval){ [ 542.030078] env[62460]: value = "task-1313405" [ 542.030078] env[62460]: _type = "Task" [ 542.030078] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 542.038907] env[62460]: DEBUG oslo_vmware.api [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': task-1313405, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 542.129839] env[62460]: DEBUG nova.network.neutron [-] [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 542.218317] env[62460]: DEBUG nova.compute.utils [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 542.226257] env[62460]: DEBUG nova.compute.manager [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 542.226463] env[62460]: DEBUG nova.network.neutron [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 542.372374] env[62460]: DEBUG nova.policy [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2fed8c51f81c4dd584efeda344d64193', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '26d2a78cc07d4ca686d7b4d3beea449a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 542.466578] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6349c113-4f3a-482a-9e14-4d0497a34898 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.477894] env[62460]: DEBUG nova.network.neutron [-] [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 542.480328] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d3e05d9-db99-4e60-bb92-d6e271a0502f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.520039] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93a441f5-51c3-4a81-930e-e10e325aa98d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.529132] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a898b397-67c6-4764-8ae8-2828038e3e4e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.545553] env[62460]: DEBUG oslo_vmware.api [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': task-1313405, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.496712} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 542.554698] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] 44a2ac83-3fff-4958-a1d3-b3884b42c211/44a2ac83-3fff-4958-a1d3-b3884b42c211.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 542.555042] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 542.555449] env[62460]: DEBUG nova.compute.provider_tree [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 542.557523] env[62460]: DEBUG nova.network.neutron [req-5811884b-3c39-4e59-8d84-238eb67cc25d req-34e5a7c9-7c76-475e-9ad3-e9ebbf55b129 service nova] [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 542.559282] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e3f4978d-07a0-42e9-adb4-7c17ea65b880 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.563819] env[62460]: DEBUG nova.network.neutron [req-082bb4f5-9d0d-40cf-b48e-54f2db013416 req-7a251895-e105-40d4-994c-c68461b669b4 service nova] [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 542.567784] env[62460]: DEBUG oslo_vmware.api [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Waiting for the task: (returnval){ [ 542.567784] env[62460]: value = "task-1313406" [ 542.567784] env[62460]: _type = "Task" [ 542.567784] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 542.578965] env[62460]: DEBUG oslo_vmware.api [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': task-1313406, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 542.636036] env[62460]: INFO nova.compute.manager [-] [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] Took 1.22 seconds to deallocate network for instance. [ 542.637840] env[62460]: DEBUG nova.compute.claims [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 542.638095] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.738711] env[62460]: DEBUG nova.compute.manager [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 542.886074] env[62460]: DEBUG nova.network.neutron [req-5811884b-3c39-4e59-8d84-238eb67cc25d req-34e5a7c9-7c76-475e-9ad3-e9ebbf55b129 service nova] [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 542.909547] env[62460]: DEBUG nova.network.neutron [req-082bb4f5-9d0d-40cf-b48e-54f2db013416 req-7a251895-e105-40d4-994c-c68461b669b4 service nova] [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 542.985687] env[62460]: INFO nova.compute.manager [-] [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] Took 1.09 seconds to deallocate network for instance. [ 542.989955] env[62460]: DEBUG nova.compute.claims [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 542.990564] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.063739] env[62460]: DEBUG nova.scheduler.client.report [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 543.086477] env[62460]: DEBUG oslo_vmware.api [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': task-1313406, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070392} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 543.086477] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 543.088208] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceffe2e4-d617-4009-9a77-1360f24d9b6b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.114493] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Reconfiguring VM instance instance-00000001 to attach disk [datastore2] 44a2ac83-3fff-4958-a1d3-b3884b42c211/44a2ac83-3fff-4958-a1d3-b3884b42c211.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 543.115558] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5c5e9769-f024-48b5-bfce-8c6e648841c9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.139039] env[62460]: DEBUG oslo_vmware.api [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Waiting for the task: (returnval){ [ 543.139039] env[62460]: value = "task-1313408" [ 543.139039] env[62460]: _type = "Task" [ 543.139039] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 543.153204] env[62460]: DEBUG oslo_vmware.api [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': task-1313408, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 543.388397] env[62460]: DEBUG oslo_concurrency.lockutils [req-5811884b-3c39-4e59-8d84-238eb67cc25d req-34e5a7c9-7c76-475e-9ad3-e9ebbf55b129 service nova] Releasing lock "refresh_cache-48e50ab0-9d4e-4c5d-96d5-55399b3a8e19" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 543.389317] env[62460]: DEBUG nova.compute.manager [req-5811884b-3c39-4e59-8d84-238eb67cc25d req-34e5a7c9-7c76-475e-9ad3-e9ebbf55b129 service nova] [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] Received event network-vif-deleted-f6eb304f-e7db-40f7-a462-0db7083694d9 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 543.394041] env[62460]: DEBUG oslo_concurrency.lockutils [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Acquiring lock "efef4327-22c8-4b59-9b26-89bcfe89bc59" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.394283] env[62460]: DEBUG oslo_concurrency.lockutils [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Lock "efef4327-22c8-4b59-9b26-89bcfe89bc59" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 543.413717] env[62460]: DEBUG oslo_concurrency.lockutils [req-082bb4f5-9d0d-40cf-b48e-54f2db013416 req-7a251895-e105-40d4-994c-c68461b669b4 service nova] Releasing lock "refresh_cache-d16440e7-aa8a-4041-a4cd-d0ea84d8c027" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 543.576322] env[62460]: DEBUG oslo_concurrency.lockutils [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.861s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 543.577344] env[62460]: ERROR nova.compute.manager [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1f7f19cc-6d5a-435a-8e7f-124093a4d283, please check neutron logs for more information. [ 543.577344] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] Traceback (most recent call last): [ 543.577344] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 543.577344] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] self.driver.spawn(context, instance, image_meta, [ 543.577344] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 543.577344] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 543.577344] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 543.577344] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] vm_ref = self.build_virtual_machine(instance, [ 543.577344] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 543.577344] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] vif_infos = vmwarevif.get_vif_info(self._session, [ 543.577344] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 543.579785] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] for vif in network_info: [ 543.579785] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 543.579785] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] return self._sync_wrapper(fn, *args, **kwargs) [ 543.579785] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 543.579785] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] self.wait() [ 543.579785] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 543.579785] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] self[:] = self._gt.wait() [ 543.579785] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 543.579785] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] return self._exit_event.wait() [ 543.579785] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 543.579785] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] result = hub.switch() [ 543.579785] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 543.579785] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] return self.greenlet.switch() [ 543.580588] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 543.580588] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] result = function(*args, **kwargs) [ 543.580588] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 543.580588] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] return func(*args, **kwargs) [ 543.580588] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 543.580588] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] raise e [ 543.580588] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 543.580588] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] nwinfo = self.network_api.allocate_for_instance( [ 543.580588] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 543.580588] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] created_port_ids = self._update_ports_for_instance( [ 543.580588] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 543.580588] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] with excutils.save_and_reraise_exception(): [ 543.580588] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 543.580988] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] self.force_reraise() [ 543.580988] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 543.580988] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] raise self.value [ 543.580988] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 543.580988] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] updated_port = self._update_port( [ 543.580988] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 543.580988] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] _ensure_no_port_binding_failure(port) [ 543.580988] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 543.580988] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] raise exception.PortBindingFailed(port_id=port['id']) [ 543.580988] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] nova.exception.PortBindingFailed: Binding failed for port 1f7f19cc-6d5a-435a-8e7f-124093a4d283, please check neutron logs for more information. [ 543.580988] env[62460]: ERROR nova.compute.manager [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] [ 543.581474] env[62460]: DEBUG nova.compute.utils [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] Binding failed for port 1f7f19cc-6d5a-435a-8e7f-124093a4d283, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 543.581474] env[62460]: DEBUG oslo_concurrency.lockutils [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 7.626s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 543.586878] env[62460]: DEBUG nova.compute.manager [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] Build of instance 82afa576-f501-4a30-b8f7-e8b6334130cb was re-scheduled: Binding failed for port 1f7f19cc-6d5a-435a-8e7f-124093a4d283, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 543.587456] env[62460]: DEBUG nova.compute.manager [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 543.588771] env[62460]: DEBUG oslo_concurrency.lockutils [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] Acquiring lock "refresh_cache-82afa576-f501-4a30-b8f7-e8b6334130cb" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 543.588771] env[62460]: DEBUG oslo_concurrency.lockutils [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] Acquired lock "refresh_cache-82afa576-f501-4a30-b8f7-e8b6334130cb" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 543.588771] env[62460]: DEBUG nova.network.neutron [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 543.655986] env[62460]: DEBUG oslo_vmware.api [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': task-1313408, 'name': ReconfigVM_Task, 'duration_secs': 0.341299} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 543.655986] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Reconfigured VM instance instance-00000001 to attach disk [datastore2] 44a2ac83-3fff-4958-a1d3-b3884b42c211/44a2ac83-3fff-4958-a1d3-b3884b42c211.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 543.655986] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-99f8054a-f8bd-4e4e-9b03-13c865d4693e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.672308] env[62460]: DEBUG oslo_vmware.api [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Waiting for the task: (returnval){ [ 543.672308] env[62460]: value = "task-1313409" [ 543.672308] env[62460]: _type = "Task" [ 543.672308] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 543.687929] env[62460]: DEBUG oslo_vmware.api [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': task-1313409, 'name': Rename_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 543.759021] env[62460]: DEBUG nova.compute.manager [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 543.803595] env[62460]: DEBUG nova.virt.hardware [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 543.803958] env[62460]: DEBUG nova.virt.hardware [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 543.804285] env[62460]: DEBUG nova.virt.hardware [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 543.804578] env[62460]: DEBUG nova.virt.hardware [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 543.804882] env[62460]: DEBUG nova.virt.hardware [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 543.805197] env[62460]: DEBUG nova.virt.hardware [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 543.805521] env[62460]: DEBUG nova.virt.hardware [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 543.805762] env[62460]: DEBUG nova.virt.hardware [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 543.806066] env[62460]: DEBUG nova.virt.hardware [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 543.806293] env[62460]: DEBUG nova.virt.hardware [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 543.806607] env[62460]: DEBUG nova.virt.hardware [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 543.808455] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ac20ca7-89cb-453e-add3-292594c608c7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.824427] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d653dfc-7a75-4a6c-bf54-fabe787859fe {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.839914] env[62460]: DEBUG nova.network.neutron [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] Successfully created port: 85fa9617-3127-47cb-8e24-46840450c4da {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 544.141021] env[62460]: DEBUG nova.network.neutron [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 544.190676] env[62460]: DEBUG oslo_vmware.api [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': task-1313409, 'name': Rename_Task, 'duration_secs': 0.164704} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 544.190959] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 544.191241] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5ab5dfca-5810-4ff7-871c-ddd4bcb7eecc {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.199821] env[62460]: DEBUG oslo_vmware.api [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Waiting for the task: (returnval){ [ 544.199821] env[62460]: value = "task-1313410" [ 544.199821] env[62460]: _type = "Task" [ 544.199821] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 544.221658] env[62460]: DEBUG oslo_vmware.api [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': task-1313410, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 544.359975] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e52dac29-0338-468e-a4ab-49e9a31a0b7c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.368643] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05c54cc8-32e4-4a90-82d7-82849b6ad003 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.409036] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecc099a5-e8f5-4770-bdfe-5e3fc61df9a6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.411392] env[62460]: DEBUG nova.network.neutron [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 544.416877] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19830b33-a119-4cfa-9977-7b64220b7c5f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.436296] env[62460]: DEBUG nova.compute.provider_tree [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 544.618856] env[62460]: DEBUG nova.compute.manager [req-d1e0ffc8-11a1-4e15-82dd-4578f5d65ed4 req-e5ff1345-340e-4f68-b802-ab33f18b1e74 service nova] [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] Received event network-vif-deleted-16f5912b-00e1-4f98-9484-4aabc70c360d {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 544.718418] env[62460]: DEBUG oslo_vmware.api [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': task-1313410, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 544.919185] env[62460]: DEBUG oslo_concurrency.lockutils [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] Releasing lock "refresh_cache-82afa576-f501-4a30-b8f7-e8b6334130cb" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 544.919185] env[62460]: DEBUG nova.compute.manager [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 544.919185] env[62460]: DEBUG nova.compute.manager [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 544.919185] env[62460]: DEBUG nova.network.neutron [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 544.940205] env[62460]: DEBUG nova.scheduler.client.report [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 544.965903] env[62460]: DEBUG nova.network.neutron [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 545.218928] env[62460]: DEBUG oslo_vmware.api [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': task-1313410, 'name': PowerOnVM_Task, 'duration_secs': 0.666072} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 545.222446] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 545.222609] env[62460]: DEBUG nova.compute.manager [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 545.223908] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15da6ce0-ad50-4b28-9a62-a1e245be6747 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.450333] env[62460]: DEBUG oslo_concurrency.lockutils [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.869s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 545.450917] env[62460]: ERROR nova.compute.manager [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e30693e4-484a-4acc-ab76-e9da7cda9a03, please check neutron logs for more information. [ 545.450917] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] Traceback (most recent call last): [ 545.450917] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 545.450917] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] self.driver.spawn(context, instance, image_meta, [ 545.450917] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 545.450917] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] self._vmops.spawn(context, instance, image_meta, injected_files, [ 545.450917] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 545.450917] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] vm_ref = self.build_virtual_machine(instance, [ 545.450917] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 545.450917] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] vif_infos = vmwarevif.get_vif_info(self._session, [ 545.450917] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 545.452577] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] for vif in network_info: [ 545.452577] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 545.452577] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] return self._sync_wrapper(fn, *args, **kwargs) [ 545.452577] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 545.452577] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] self.wait() [ 545.452577] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 545.452577] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] self[:] = self._gt.wait() [ 545.452577] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 545.452577] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] return self._exit_event.wait() [ 545.452577] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 545.452577] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] result = hub.switch() [ 545.452577] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 545.452577] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] return self.greenlet.switch() [ 545.453108] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 545.453108] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] result = function(*args, **kwargs) [ 545.453108] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 545.453108] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] return func(*args, **kwargs) [ 545.453108] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 545.453108] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] raise e [ 545.453108] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 545.453108] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] nwinfo = self.network_api.allocate_for_instance( [ 545.453108] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 545.453108] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] created_port_ids = self._update_ports_for_instance( [ 545.453108] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 545.453108] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] with excutils.save_and_reraise_exception(): [ 545.453108] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 545.453622] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] self.force_reraise() [ 545.453622] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 545.453622] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] raise self.value [ 545.453622] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 545.453622] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] updated_port = self._update_port( [ 545.453622] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 545.453622] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] _ensure_no_port_binding_failure(port) [ 545.453622] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 545.453622] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] raise exception.PortBindingFailed(port_id=port['id']) [ 545.453622] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] nova.exception.PortBindingFailed: Binding failed for port e30693e4-484a-4acc-ab76-e9da7cda9a03, please check neutron logs for more information. [ 545.453622] env[62460]: ERROR nova.compute.manager [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] [ 545.454054] env[62460]: DEBUG nova.compute.utils [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] Binding failed for port e30693e4-484a-4acc-ab76-e9da7cda9a03, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 545.454054] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 8.274s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 545.462024] env[62460]: DEBUG nova.compute.manager [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] Build of instance 61d0d996-78ad-4b20-a7f4-b8a8f130d969 was re-scheduled: Binding failed for port e30693e4-484a-4acc-ab76-e9da7cda9a03, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 545.462024] env[62460]: DEBUG nova.compute.manager [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 545.462024] env[62460]: DEBUG oslo_concurrency.lockutils [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Acquiring lock "refresh_cache-61d0d996-78ad-4b20-a7f4-b8a8f130d969" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 545.462024] env[62460]: DEBUG oslo_concurrency.lockutils [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Acquired lock "refresh_cache-61d0d996-78ad-4b20-a7f4-b8a8f130d969" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 545.462390] env[62460]: DEBUG nova.network.neutron [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 545.470173] env[62460]: DEBUG nova.network.neutron [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 545.751250] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.974096] env[62460]: INFO nova.compute.manager [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] [instance: 82afa576-f501-4a30-b8f7-e8b6334130cb] Took 1.06 seconds to deallocate network for instance. [ 546.001357] env[62460]: DEBUG nova.network.neutron [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 546.106629] env[62460]: DEBUG nova.network.neutron [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 546.172316] env[62460]: DEBUG oslo_concurrency.lockutils [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Acquiring lock "8a2699f3-3cc4-4ec3-899a-216e09ad086a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 546.172575] env[62460]: DEBUG oslo_concurrency.lockutils [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Lock "8a2699f3-3cc4-4ec3-899a-216e09ad086a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 546.222222] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b1f925a-f1c2-4f3c-ac49-45e4c73e8f56 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.231495] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5128e37-d120-418e-bf9d-341e8abeb936 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.265675] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a95cfd4-b240-496b-bbe7-359fa6bb9d2a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.274268] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae702979-3c3e-45d6-a0da-f9885b7c42a9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.290051] env[62460]: DEBUG nova.compute.provider_tree [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 546.611460] env[62460]: DEBUG oslo_concurrency.lockutils [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Releasing lock "refresh_cache-61d0d996-78ad-4b20-a7f4-b8a8f130d969" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 546.611726] env[62460]: DEBUG nova.compute.manager [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 546.611864] env[62460]: DEBUG nova.compute.manager [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 546.612042] env[62460]: DEBUG nova.network.neutron [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 546.640960] env[62460]: DEBUG nova.network.neutron [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 546.794326] env[62460]: DEBUG nova.scheduler.client.report [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 547.022777] env[62460]: INFO nova.scheduler.client.report [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] Deleted allocations for instance 82afa576-f501-4a30-b8f7-e8b6334130cb [ 547.113029] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] Acquiring lock "a3f7edbe-4bba-4fff-9e62-99b7a85f971b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.113029] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] Lock "a3f7edbe-4bba-4fff-9e62-99b7a85f971b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.147319] env[62460]: DEBUG nova.network.neutron [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 547.300443] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.847s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 547.301116] env[62460]: ERROR nova.compute.manager [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 228ca897-3a3c-4e88-9838-451cdc94d456, please check neutron logs for more information. [ 547.301116] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] Traceback (most recent call last): [ 547.301116] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 547.301116] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] self.driver.spawn(context, instance, image_meta, [ 547.301116] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 547.301116] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 547.301116] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 547.301116] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] vm_ref = self.build_virtual_machine(instance, [ 547.301116] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 547.301116] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] vif_infos = vmwarevif.get_vif_info(self._session, [ 547.301116] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 547.304330] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] for vif in network_info: [ 547.304330] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 547.304330] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] return self._sync_wrapper(fn, *args, **kwargs) [ 547.304330] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 547.304330] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] self.wait() [ 547.304330] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 547.304330] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] self[:] = self._gt.wait() [ 547.304330] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 547.304330] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] return self._exit_event.wait() [ 547.304330] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 547.304330] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] result = hub.switch() [ 547.304330] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 547.304330] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] return self.greenlet.switch() [ 547.304818] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 547.304818] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] result = function(*args, **kwargs) [ 547.304818] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 547.304818] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] return func(*args, **kwargs) [ 547.304818] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 547.304818] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] raise e [ 547.304818] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 547.304818] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] nwinfo = self.network_api.allocate_for_instance( [ 547.304818] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 547.304818] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] created_port_ids = self._update_ports_for_instance( [ 547.304818] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 547.304818] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] with excutils.save_and_reraise_exception(): [ 547.304818] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 547.305183] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] self.force_reraise() [ 547.305183] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 547.305183] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] raise self.value [ 547.305183] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 547.305183] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] updated_port = self._update_port( [ 547.305183] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 547.305183] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] _ensure_no_port_binding_failure(port) [ 547.305183] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 547.305183] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] raise exception.PortBindingFailed(port_id=port['id']) [ 547.305183] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] nova.exception.PortBindingFailed: Binding failed for port 228ca897-3a3c-4e88-9838-451cdc94d456, please check neutron logs for more information. [ 547.305183] env[62460]: ERROR nova.compute.manager [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] [ 547.305481] env[62460]: DEBUG nova.compute.utils [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] Binding failed for port 228ca897-3a3c-4e88-9838-451cdc94d456, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 547.306397] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.156s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.308348] env[62460]: INFO nova.compute.claims [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 547.319319] env[62460]: DEBUG nova.compute.manager [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] Build of instance 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f was re-scheduled: Binding failed for port 228ca897-3a3c-4e88-9838-451cdc94d456, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 547.319319] env[62460]: DEBUG nova.compute.manager [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 547.319515] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Acquiring lock "refresh_cache-2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 547.319670] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Acquired lock "refresh_cache-2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 547.319840] env[62460]: DEBUG nova.network.neutron [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 547.321740] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "b51ddc3a-de97-4d51-adc9-e92b4fb0a150" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.321952] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "b51ddc3a-de97-4d51-adc9-e92b4fb0a150" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.482034] env[62460]: ERROR nova.compute.manager [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 85fa9617-3127-47cb-8e24-46840450c4da, please check neutron logs for more information. [ 547.482034] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 547.482034] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 547.482034] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 547.482034] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 547.482034] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 547.482034] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 547.482034] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 547.482034] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 547.482034] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 547.482034] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 547.482034] env[62460]: ERROR nova.compute.manager raise self.value [ 547.482034] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 547.482034] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 547.482034] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 547.482034] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 547.482546] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 547.482546] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 547.482546] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 85fa9617-3127-47cb-8e24-46840450c4da, please check neutron logs for more information. [ 547.482546] env[62460]: ERROR nova.compute.manager [ 547.483370] env[62460]: Traceback (most recent call last): [ 547.483448] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 547.483448] env[62460]: listener.cb(fileno) [ 547.483559] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 547.483559] env[62460]: result = function(*args, **kwargs) [ 547.483629] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 547.483629] env[62460]: return func(*args, **kwargs) [ 547.483692] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 547.483692] env[62460]: raise e [ 547.483758] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 547.483758] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 547.483822] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 547.483822] env[62460]: created_port_ids = self._update_ports_for_instance( [ 547.483884] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 547.483884] env[62460]: with excutils.save_and_reraise_exception(): [ 547.483954] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 547.483954] env[62460]: self.force_reraise() [ 547.484030] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 547.484030] env[62460]: raise self.value [ 547.484093] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 547.484093] env[62460]: updated_port = self._update_port( [ 547.484156] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 547.484156] env[62460]: _ensure_no_port_binding_failure(port) [ 547.484269] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 547.484269] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 547.484346] env[62460]: nova.exception.PortBindingFailed: Binding failed for port 85fa9617-3127-47cb-8e24-46840450c4da, please check neutron logs for more information. [ 547.484390] env[62460]: Removing descriptor: 18 [ 547.485245] env[62460]: ERROR nova.compute.manager [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 85fa9617-3127-47cb-8e24-46840450c4da, please check neutron logs for more information. [ 547.485245] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] Traceback (most recent call last): [ 547.485245] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 547.485245] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] yield resources [ 547.485245] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 547.485245] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] self.driver.spawn(context, instance, image_meta, [ 547.485245] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 547.485245] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 547.485245] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 547.485245] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] vm_ref = self.build_virtual_machine(instance, [ 547.485245] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 547.485602] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] vif_infos = vmwarevif.get_vif_info(self._session, [ 547.485602] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 547.485602] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] for vif in network_info: [ 547.485602] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 547.485602] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] return self._sync_wrapper(fn, *args, **kwargs) [ 547.485602] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 547.485602] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] self.wait() [ 547.485602] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 547.485602] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] self[:] = self._gt.wait() [ 547.485602] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 547.485602] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] return self._exit_event.wait() [ 547.485602] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 547.485602] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] result = hub.switch() [ 547.485930] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 547.485930] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] return self.greenlet.switch() [ 547.485930] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 547.485930] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] result = function(*args, **kwargs) [ 547.485930] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 547.485930] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] return func(*args, **kwargs) [ 547.485930] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 547.485930] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] raise e [ 547.485930] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 547.485930] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] nwinfo = self.network_api.allocate_for_instance( [ 547.485930] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 547.485930] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] created_port_ids = self._update_ports_for_instance( [ 547.485930] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 547.486387] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] with excutils.save_and_reraise_exception(): [ 547.486387] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 547.486387] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] self.force_reraise() [ 547.486387] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 547.486387] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] raise self.value [ 547.486387] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 547.486387] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] updated_port = self._update_port( [ 547.486387] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 547.486387] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] _ensure_no_port_binding_failure(port) [ 547.486387] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 547.486387] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] raise exception.PortBindingFailed(port_id=port['id']) [ 547.486387] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] nova.exception.PortBindingFailed: Binding failed for port 85fa9617-3127-47cb-8e24-46840450c4da, please check neutron logs for more information. [ 547.486387] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] [ 547.488022] env[62460]: INFO nova.compute.manager [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] Terminating instance [ 547.491924] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] Acquiring lock "refresh_cache-72714a2e-a95d-495e-b163-f5168148fcdf" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 547.491924] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] Acquired lock "refresh_cache-72714a2e-a95d-495e-b163-f5168148fcdf" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 547.491924] env[62460]: DEBUG nova.network.neutron [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 547.541041] env[62460]: DEBUG oslo_concurrency.lockutils [None req-42a658d8-7567-44db-b7ab-8a69192fd38c tempest-ServersAdminNegativeTestJSON-1676025269 tempest-ServersAdminNegativeTestJSON-1676025269-project-member] Lock "82afa576-f501-4a30-b8f7-e8b6334130cb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.235s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 547.653122] env[62460]: INFO nova.compute.manager [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 61d0d996-78ad-4b20-a7f4-b8a8f130d969] Took 1.04 seconds to deallocate network for instance. [ 547.674860] env[62460]: DEBUG oslo_concurrency.lockutils [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Acquiring lock "6ea9a22a-d6ca-4e14-b83e-97074d4cb531" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.675488] env[62460]: DEBUG oslo_concurrency.lockutils [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Lock "6ea9a22a-d6ca-4e14-b83e-97074d4cb531" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.855976] env[62460]: DEBUG nova.network.neutron [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 548.046209] env[62460]: DEBUG nova.network.neutron [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 548.049533] env[62460]: DEBUG nova.compute.manager [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] [instance: 00544d2a-1a15-4347-abe4-3641b5bf0cfe] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 548.131219] env[62460]: DEBUG nova.network.neutron [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 548.196337] env[62460]: DEBUG nova.compute.manager [req-85f105da-c57b-4dac-9bf7-3a25a58d1346 req-a2641ae3-b1eb-4507-b9c6-04405192ad78 service nova] [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] Received event network-changed-85fa9617-3127-47cb-8e24-46840450c4da {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 548.196559] env[62460]: DEBUG nova.compute.manager [req-85f105da-c57b-4dac-9bf7-3a25a58d1346 req-a2641ae3-b1eb-4507-b9c6-04405192ad78 service nova] [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] Refreshing instance network info cache due to event network-changed-85fa9617-3127-47cb-8e24-46840450c4da. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 548.196913] env[62460]: DEBUG oslo_concurrency.lockutils [req-85f105da-c57b-4dac-9bf7-3a25a58d1346 req-a2641ae3-b1eb-4507-b9c6-04405192ad78 service nova] Acquiring lock "refresh_cache-72714a2e-a95d-495e-b163-f5168148fcdf" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 548.369064] env[62460]: DEBUG nova.network.neutron [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 548.580900] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.615795] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93ed421d-c7e2-4c10-95da-3df1f06234de {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.628458] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fa72524-1bb8-4259-9702-38936d3c9278 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.636338] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Releasing lock "refresh_cache-2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 548.636338] env[62460]: DEBUG nova.compute.manager [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 548.638291] env[62460]: DEBUG nova.compute.manager [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 548.640815] env[62460]: DEBUG nova.network.neutron [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 548.683723] env[62460]: DEBUG nova.network.neutron [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 548.687023] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51c2391b-137c-4ef5-8ce6-6fc09754f559 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.696179] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9244e68-7d28-4e2c-a7d9-e1d0653f9c51 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.713421] env[62460]: DEBUG nova.compute.provider_tree [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 548.716640] env[62460]: INFO nova.scheduler.client.report [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Deleted allocations for instance 61d0d996-78ad-4b20-a7f4-b8a8f130d969 [ 548.875215] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] Releasing lock "refresh_cache-72714a2e-a95d-495e-b163-f5168148fcdf" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 548.875676] env[62460]: DEBUG nova.compute.manager [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 548.875868] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 548.876212] env[62460]: DEBUG oslo_concurrency.lockutils [req-85f105da-c57b-4dac-9bf7-3a25a58d1346 req-a2641ae3-b1eb-4507-b9c6-04405192ad78 service nova] Acquired lock "refresh_cache-72714a2e-a95d-495e-b163-f5168148fcdf" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 548.876391] env[62460]: DEBUG nova.network.neutron [req-85f105da-c57b-4dac-9bf7-3a25a58d1346 req-a2641ae3-b1eb-4507-b9c6-04405192ad78 service nova] [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] Refreshing network info cache for port 85fa9617-3127-47cb-8e24-46840450c4da {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 548.877565] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-58e8aaf3-adbe-496b-b29d-64d9767dc373 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.893035] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8de7c1c8-002d-4200-b1a1-19853c4674e9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.922568] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 72714a2e-a95d-495e-b163-f5168148fcdf could not be found. [ 548.922825] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 548.923018] env[62460]: INFO nova.compute.manager [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] Took 0.05 seconds to destroy the instance on the hypervisor. [ 548.923279] env[62460]: DEBUG oslo.service.loopingcall [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 548.923527] env[62460]: DEBUG nova.compute.manager [-] [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 548.923630] env[62460]: DEBUG nova.network.neutron [-] [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 548.946067] env[62460]: DEBUG nova.network.neutron [-] [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 549.193023] env[62460]: DEBUG nova.network.neutron [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 549.213022] env[62460]: DEBUG oslo_concurrency.lockutils [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] Acquiring lock "291beb16-db0e-4eb6-a224-2acfc14d6d77" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.213022] env[62460]: DEBUG oslo_concurrency.lockutils [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] Lock "291beb16-db0e-4eb6-a224-2acfc14d6d77" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.229557] env[62460]: DEBUG nova.scheduler.client.report [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 549.234522] env[62460]: DEBUG oslo_concurrency.lockutils [None req-82541b1f-fba1-4d6a-8514-54a228504215 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Lock "61d0d996-78ad-4b20-a7f4-b8a8f130d969" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.716s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 549.411735] env[62460]: DEBUG nova.network.neutron [req-85f105da-c57b-4dac-9bf7-3a25a58d1346 req-a2641ae3-b1eb-4507-b9c6-04405192ad78 service nova] [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 549.450093] env[62460]: DEBUG nova.network.neutron [-] [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 549.695775] env[62460]: INFO nova.compute.manager [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f] Took 1.06 seconds to deallocate network for instance. [ 549.726788] env[62460]: DEBUG oslo_concurrency.lockutils [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Acquiring lock "9b8ff88e-9e13-42b2-92b0-13af6873c916" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.726788] env[62460]: DEBUG oslo_concurrency.lockutils [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Lock "9b8ff88e-9e13-42b2-92b0-13af6873c916" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.739179] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.432s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 549.739432] env[62460]: DEBUG nova.compute.manager [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 549.741964] env[62460]: DEBUG nova.compute.manager [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 549.747886] env[62460]: DEBUG oslo_concurrency.lockutils [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 10.612s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.772787] env[62460]: DEBUG nova.network.neutron [req-85f105da-c57b-4dac-9bf7-3a25a58d1346 req-a2641ae3-b1eb-4507-b9c6-04405192ad78 service nova] [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 549.954020] env[62460]: INFO nova.compute.manager [-] [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] Took 1.03 seconds to deallocate network for instance. [ 549.958158] env[62460]: DEBUG nova.compute.claims [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 549.958158] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.249918] env[62460]: DEBUG nova.compute.utils [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 550.251603] env[62460]: DEBUG nova.compute.manager [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 550.251603] env[62460]: DEBUG nova.network.neutron [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 550.275627] env[62460]: DEBUG oslo_concurrency.lockutils [req-85f105da-c57b-4dac-9bf7-3a25a58d1346 req-a2641ae3-b1eb-4507-b9c6-04405192ad78 service nova] Releasing lock "refresh_cache-72714a2e-a95d-495e-b163-f5168148fcdf" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 550.290263] env[62460]: DEBUG oslo_concurrency.lockutils [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.291541] env[62460]: INFO nova.compute.manager [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Rebuilding instance [ 550.344049] env[62460]: DEBUG nova.compute.manager [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 550.344903] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ca3c022-2b9d-483e-83c3-c052ad3b8653 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.349224] env[62460]: DEBUG nova.policy [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7684644b29ad4d709309cadf8c154449', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '19e53e9f113e4465924f75c1f9a39008', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 550.535412] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-525c4afd-e1f3-4ac5-814f-87acb78b09b6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.543836] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68bf6b78-ac32-4049-9333-9c4e97494013 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.577252] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd1bf97e-5533-4049-a36e-5c1c4cd4ce8c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.586431] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03a608bc-373e-4663-9bb9-14f00b02c3dc {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.601831] env[62460]: DEBUG nova.compute.provider_tree [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 550.739416] env[62460]: INFO nova.scheduler.client.report [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Deleted allocations for instance 2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f [ 550.757406] env[62460]: DEBUG nova.compute.manager [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 550.815704] env[62460]: DEBUG nova.network.neutron [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] Successfully created port: 0de37ccf-02a2-4a86-9bb6-d49a72ab3f31 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 550.866109] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 550.866469] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6258c6c5-9606-42bd-aa4e-02f106c3bb99 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.877014] env[62460]: DEBUG oslo_vmware.api [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Waiting for the task: (returnval){ [ 550.877014] env[62460]: value = "task-1313414" [ 550.877014] env[62460]: _type = "Task" [ 550.877014] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 550.886916] env[62460]: DEBUG oslo_vmware.api [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Task: {'id': task-1313414, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 551.108222] env[62460]: DEBUG nova.scheduler.client.report [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 551.249462] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b7cd36a5-7dd3-48bc-a845-7f34ba19d880 tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Lock "2ad87bc0-a6c1-4a63-bb56-8cea825ffc4f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.906s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 551.253075] env[62460]: DEBUG nova.network.neutron [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] Successfully created port: ce18fd45-4f7b-42f3-9ec8-b2cb9e37d0d2 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 551.389493] env[62460]: DEBUG oslo_vmware.api [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Task: {'id': task-1313414, 'name': PowerOffVM_Task, 'duration_secs': 0.125542} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 551.389493] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 551.389493] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 551.390678] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad649ef2-6e83-4368-9b8c-623ba0cb188e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.401149] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 551.401908] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-352a8c30-e952-490c-9027-56c13cc59737 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.436108] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 551.436357] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Deleting contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 551.436541] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Deleting the datastore file [datastore2] 44a2ac83-3fff-4958-a1d3-b3884b42c211 {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 551.436809] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cf96306a-34e4-4ce4-bb05-95606561353f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.446677] env[62460]: DEBUG oslo_vmware.api [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Waiting for the task: (returnval){ [ 551.446677] env[62460]: value = "task-1313416" [ 551.446677] env[62460]: _type = "Task" [ 551.446677] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 551.460360] env[62460]: DEBUG oslo_vmware.api [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Task: {'id': task-1313416, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 551.614435] env[62460]: DEBUG oslo_concurrency.lockutils [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.866s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 551.616191] env[62460]: ERROR nova.compute.manager [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] [instance: 7da5a822-6681-4dc6-b637-926ce425a042] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d8646801-5db8-41d8-a53c-f24e8f20d19a, please check neutron logs for more information. [ 551.616191] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] Traceback (most recent call last): [ 551.616191] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 551.616191] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] self.driver.spawn(context, instance, image_meta, [ 551.616191] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 551.616191] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] self._vmops.spawn(context, instance, image_meta, injected_files, [ 551.616191] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 551.616191] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] vm_ref = self.build_virtual_machine(instance, [ 551.616191] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 551.616191] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] vif_infos = vmwarevif.get_vif_info(self._session, [ 551.616191] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 551.616624] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] for vif in network_info: [ 551.616624] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 551.616624] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] return self._sync_wrapper(fn, *args, **kwargs) [ 551.616624] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 551.616624] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] self.wait() [ 551.616624] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 551.616624] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] self[:] = self._gt.wait() [ 551.616624] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 551.616624] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] return self._exit_event.wait() [ 551.616624] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 551.616624] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] result = hub.switch() [ 551.616624] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 551.616624] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] return self.greenlet.switch() [ 551.616938] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 551.616938] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] result = function(*args, **kwargs) [ 551.616938] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 551.616938] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] return func(*args, **kwargs) [ 551.616938] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 551.616938] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] raise e [ 551.616938] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 551.616938] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] nwinfo = self.network_api.allocate_for_instance( [ 551.616938] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 551.616938] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] created_port_ids = self._update_ports_for_instance( [ 551.616938] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 551.616938] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] with excutils.save_and_reraise_exception(): [ 551.616938] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 551.617508] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] self.force_reraise() [ 551.617508] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 551.617508] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] raise self.value [ 551.617508] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 551.617508] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] updated_port = self._update_port( [ 551.617508] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 551.617508] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] _ensure_no_port_binding_failure(port) [ 551.617508] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 551.617508] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] raise exception.PortBindingFailed(port_id=port['id']) [ 551.617508] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] nova.exception.PortBindingFailed: Binding failed for port d8646801-5db8-41d8-a53c-f24e8f20d19a, please check neutron logs for more information. [ 551.617508] env[62460]: ERROR nova.compute.manager [instance: 7da5a822-6681-4dc6-b637-926ce425a042] [ 551.618014] env[62460]: DEBUG nova.compute.utils [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] [instance: 7da5a822-6681-4dc6-b637-926ce425a042] Binding failed for port d8646801-5db8-41d8-a53c-f24e8f20d19a, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 551.618199] env[62460]: DEBUG nova.compute.manager [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] [instance: 7da5a822-6681-4dc6-b637-926ce425a042] Build of instance 7da5a822-6681-4dc6-b637-926ce425a042 was re-scheduled: Binding failed for port d8646801-5db8-41d8-a53c-f24e8f20d19a, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 551.618634] env[62460]: DEBUG nova.compute.manager [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] [instance: 7da5a822-6681-4dc6-b637-926ce425a042] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 551.618865] env[62460]: DEBUG oslo_concurrency.lockutils [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] Acquiring lock "refresh_cache-7da5a822-6681-4dc6-b637-926ce425a042" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 551.619022] env[62460]: DEBUG oslo_concurrency.lockutils [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] Acquired lock "refresh_cache-7da5a822-6681-4dc6-b637-926ce425a042" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 551.619189] env[62460]: DEBUG nova.network.neutron [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] [instance: 7da5a822-6681-4dc6-b637-926ce425a042] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 551.621179] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.093s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.625018] env[62460]: INFO nova.compute.claims [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 551.715262] env[62460]: DEBUG nova.network.neutron [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] Successfully created port: 5ee9c0e0-e22f-4191-889a-8eb199831400 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 551.755351] env[62460]: DEBUG nova.compute.manager [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 551.770508] env[62460]: DEBUG nova.compute.manager [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 551.806060] env[62460]: DEBUG nova.compute.manager [req-9671e2b8-70a8-4712-99e9-01288873a0ba req-4615ba5d-72c3-4efb-9de8-6c03939c085f service nova] [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] Received event network-vif-deleted-85fa9617-3127-47cb-8e24-46840450c4da {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 551.826450] env[62460]: DEBUG nova.virt.hardware [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 551.826596] env[62460]: DEBUG nova.virt.hardware [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 551.827880] env[62460]: DEBUG nova.virt.hardware [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 551.827880] env[62460]: DEBUG nova.virt.hardware [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 551.829474] env[62460]: DEBUG nova.virt.hardware [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 551.829474] env[62460]: DEBUG nova.virt.hardware [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 551.829474] env[62460]: DEBUG nova.virt.hardware [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 551.829474] env[62460]: DEBUG nova.virt.hardware [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 551.829734] env[62460]: DEBUG nova.virt.hardware [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 551.830466] env[62460]: DEBUG nova.virt.hardware [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 551.830716] env[62460]: DEBUG nova.virt.hardware [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 551.832026] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b3cb52f-6d04-44d9-9ebc-49ce50fa7013 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.847388] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c31a10ee-7dc5-4b8b-89af-494c35b4cde6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.967599] env[62460]: DEBUG oslo_vmware.api [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Task: {'id': task-1313416, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.117977} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 551.968081] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 551.970015] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Deleted contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 551.970015] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 552.150484] env[62460]: DEBUG nova.network.neutron [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] [instance: 7da5a822-6681-4dc6-b637-926ce425a042] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 552.285864] env[62460]: DEBUG oslo_concurrency.lockutils [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.292684] env[62460]: DEBUG nova.network.neutron [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] [instance: 7da5a822-6681-4dc6-b637-926ce425a042] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 552.433885] env[62460]: DEBUG oslo_concurrency.lockutils [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] Acquiring lock "a8d7055e-e7a3-470a-ae96-f89753afa9df" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.434136] env[62460]: DEBUG oslo_concurrency.lockutils [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] Lock "a8d7055e-e7a3-470a-ae96-f89753afa9df" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 552.797947] env[62460]: DEBUG oslo_concurrency.lockutils [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] Releasing lock "refresh_cache-7da5a822-6681-4dc6-b637-926ce425a042" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 552.797947] env[62460]: DEBUG nova.compute.manager [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 552.797947] env[62460]: DEBUG nova.compute.manager [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] [instance: 7da5a822-6681-4dc6-b637-926ce425a042] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 552.797947] env[62460]: DEBUG nova.network.neutron [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] [instance: 7da5a822-6681-4dc6-b637-926ce425a042] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 552.827115] env[62460]: DEBUG nova.network.neutron [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] [instance: 7da5a822-6681-4dc6-b637-926ce425a042] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 552.992440] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a1be58f-e929-4eac-ae20-ff939c91f733 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.001476] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-898f0ce3-e574-4704-b7ff-cb319486518e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.039394] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-413a19bd-963f-4d93-9fca-7aac71f5dc8e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.046040] env[62460]: DEBUG nova.virt.hardware [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 553.046040] env[62460]: DEBUG nova.virt.hardware [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 553.046312] env[62460]: DEBUG nova.virt.hardware [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 553.046393] env[62460]: DEBUG nova.virt.hardware [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 553.046581] env[62460]: DEBUG nova.virt.hardware [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 553.046739] env[62460]: DEBUG nova.virt.hardware [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 553.046822] env[62460]: DEBUG nova.virt.hardware [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 553.046974] env[62460]: DEBUG nova.virt.hardware [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 553.047158] env[62460]: DEBUG nova.virt.hardware [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 553.047336] env[62460]: DEBUG nova.virt.hardware [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 553.047489] env[62460]: DEBUG nova.virt.hardware [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 553.048399] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-514b369a-2af7-4740-861e-463d19387806 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.056070] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44150a65-d81a-4394-8d1e-8458f2a57de3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.063531] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56f08f65-dde7-47bf-b3d3-7e8e3ec5c28f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.076446] env[62460]: DEBUG nova.compute.provider_tree [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 553.091095] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Instance VIF info [] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 553.097179] env[62460]: DEBUG oslo.service.loopingcall [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 553.098182] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 553.098414] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1e2b61a4-4559-4c95-b862-1244e83b3655 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.120321] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 553.120321] env[62460]: value = "task-1313417" [ 553.120321] env[62460]: _type = "Task" [ 553.120321] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 553.129625] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313417, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 553.331730] env[62460]: DEBUG nova.network.neutron [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] [instance: 7da5a822-6681-4dc6-b637-926ce425a042] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 553.344143] env[62460]: ERROR nova.compute.manager [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0de37ccf-02a2-4a86-9bb6-d49a72ab3f31, please check neutron logs for more information. [ 553.344143] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 553.344143] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 553.344143] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 553.344143] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 553.344143] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 553.344143] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 553.344143] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 553.344143] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 553.344143] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 553.344143] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 553.344143] env[62460]: ERROR nova.compute.manager raise self.value [ 553.344143] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 553.344143] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 553.344143] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 553.344143] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 553.344904] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 553.344904] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 553.344904] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0de37ccf-02a2-4a86-9bb6-d49a72ab3f31, please check neutron logs for more information. [ 553.344904] env[62460]: ERROR nova.compute.manager [ 553.344904] env[62460]: Traceback (most recent call last): [ 553.344904] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 553.344904] env[62460]: listener.cb(fileno) [ 553.344904] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 553.344904] env[62460]: result = function(*args, **kwargs) [ 553.344904] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 553.344904] env[62460]: return func(*args, **kwargs) [ 553.344904] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 553.344904] env[62460]: raise e [ 553.344904] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 553.344904] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 553.344904] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 553.344904] env[62460]: created_port_ids = self._update_ports_for_instance( [ 553.344904] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 553.344904] env[62460]: with excutils.save_and_reraise_exception(): [ 553.344904] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 553.344904] env[62460]: self.force_reraise() [ 553.344904] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 553.344904] env[62460]: raise self.value [ 553.344904] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 553.344904] env[62460]: updated_port = self._update_port( [ 553.344904] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 553.344904] env[62460]: _ensure_no_port_binding_failure(port) [ 553.344904] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 553.344904] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 553.345666] env[62460]: nova.exception.PortBindingFailed: Binding failed for port 0de37ccf-02a2-4a86-9bb6-d49a72ab3f31, please check neutron logs for more information. [ 553.345666] env[62460]: Removing descriptor: 17 [ 553.345666] env[62460]: ERROR nova.compute.manager [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0de37ccf-02a2-4a86-9bb6-d49a72ab3f31, please check neutron logs for more information. [ 553.345666] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] Traceback (most recent call last): [ 553.345666] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 553.345666] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] yield resources [ 553.345666] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 553.345666] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] self.driver.spawn(context, instance, image_meta, [ 553.345666] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 553.345666] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 553.345666] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 553.345666] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] vm_ref = self.build_virtual_machine(instance, [ 553.346079] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 553.346079] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] vif_infos = vmwarevif.get_vif_info(self._session, [ 553.346079] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 553.346079] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] for vif in network_info: [ 553.346079] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 553.346079] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] return self._sync_wrapper(fn, *args, **kwargs) [ 553.346079] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 553.346079] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] self.wait() [ 553.346079] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 553.346079] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] self[:] = self._gt.wait() [ 553.346079] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 553.346079] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] return self._exit_event.wait() [ 553.346079] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 553.346598] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] result = hub.switch() [ 553.346598] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 553.346598] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] return self.greenlet.switch() [ 553.346598] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 553.346598] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] result = function(*args, **kwargs) [ 553.346598] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 553.346598] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] return func(*args, **kwargs) [ 553.346598] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 553.346598] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] raise e [ 553.346598] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 553.346598] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] nwinfo = self.network_api.allocate_for_instance( [ 553.346598] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 553.346598] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] created_port_ids = self._update_ports_for_instance( [ 553.346919] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 553.346919] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] with excutils.save_and_reraise_exception(): [ 553.346919] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 553.346919] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] self.force_reraise() [ 553.346919] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 553.346919] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] raise self.value [ 553.346919] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 553.346919] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] updated_port = self._update_port( [ 553.346919] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 553.346919] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] _ensure_no_port_binding_failure(port) [ 553.346919] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 553.346919] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] raise exception.PortBindingFailed(port_id=port['id']) [ 553.347236] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] nova.exception.PortBindingFailed: Binding failed for port 0de37ccf-02a2-4a86-9bb6-d49a72ab3f31, please check neutron logs for more information. [ 553.347236] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] [ 553.347236] env[62460]: INFO nova.compute.manager [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] Terminating instance [ 553.347953] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Acquiring lock "refresh_cache-ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 553.348167] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Acquired lock "refresh_cache-ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 553.348338] env[62460]: DEBUG nova.network.neutron [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 553.580534] env[62460]: DEBUG nova.scheduler.client.report [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 553.635041] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313417, 'name': CreateVM_Task} progress is 99%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 553.836495] env[62460]: INFO nova.compute.manager [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] [instance: 7da5a822-6681-4dc6-b637-926ce425a042] Took 1.04 seconds to deallocate network for instance. [ 553.892548] env[62460]: DEBUG nova.network.neutron [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 553.987988] env[62460]: DEBUG nova.network.neutron [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 554.091595] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.471s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 554.092651] env[62460]: DEBUG nova.compute.manager [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 554.098402] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.560s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 554.135010] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313417, 'name': CreateVM_Task, 'duration_secs': 0.607104} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 554.135269] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 554.135784] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 554.136015] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 554.136403] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Acquired external semaphore "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 554.136706] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5bfe5a99-a126-4097-a509-05e38078f058 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.143132] env[62460]: DEBUG oslo_vmware.api [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Waiting for the task: (returnval){ [ 554.143132] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]520967ff-ed71-6240-729d-f854b784a946" [ 554.143132] env[62460]: _type = "Task" [ 554.143132] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 554.154846] env[62460]: DEBUG oslo_vmware.api [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]520967ff-ed71-6240-729d-f854b784a946, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 554.251440] env[62460]: DEBUG nova.compute.manager [req-25bfc2ce-e253-4a67-806f-272848d597af req-e74f4465-772e-4174-9608-9b1d9e38af8b service nova] [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] Received event network-changed-0de37ccf-02a2-4a86-9bb6-d49a72ab3f31 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 554.251761] env[62460]: DEBUG nova.compute.manager [req-25bfc2ce-e253-4a67-806f-272848d597af req-e74f4465-772e-4174-9608-9b1d9e38af8b service nova] [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] Refreshing instance network info cache due to event network-changed-0de37ccf-02a2-4a86-9bb6-d49a72ab3f31. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 554.251990] env[62460]: DEBUG oslo_concurrency.lockutils [req-25bfc2ce-e253-4a67-806f-272848d597af req-e74f4465-772e-4174-9608-9b1d9e38af8b service nova] Acquiring lock "refresh_cache-ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 554.493185] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Releasing lock "refresh_cache-ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 554.493185] env[62460]: DEBUG nova.compute.manager [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 554.493185] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 554.493185] env[62460]: DEBUG oslo_concurrency.lockutils [req-25bfc2ce-e253-4a67-806f-272848d597af req-e74f4465-772e-4174-9608-9b1d9e38af8b service nova] Acquired lock "refresh_cache-ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 554.493185] env[62460]: DEBUG nova.network.neutron [req-25bfc2ce-e253-4a67-806f-272848d597af req-e74f4465-772e-4174-9608-9b1d9e38af8b service nova] [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] Refreshing network info cache for port 0de37ccf-02a2-4a86-9bb6-d49a72ab3f31 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 554.493855] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9566ddfc-15a5-4963-8728-6947fe5f979d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.505904] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d287372-8658-4ff0-b8b4-0b59f29d0a84 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.534512] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] Acquiring lock "a2747204-95c5-4200-8742-9ec39b4368c8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.534512] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] Lock "a2747204-95c5-4200-8742-9ec39b4368c8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 554.539033] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3 could not be found. [ 554.540075] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 554.540075] env[62460]: INFO nova.compute.manager [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] Took 0.05 seconds to destroy the instance on the hypervisor. [ 554.540075] env[62460]: DEBUG oslo.service.loopingcall [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 554.540075] env[62460]: DEBUG nova.compute.manager [-] [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 554.540075] env[62460]: DEBUG nova.network.neutron [-] [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 554.606235] env[62460]: DEBUG nova.compute.utils [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 554.609286] env[62460]: DEBUG nova.compute.manager [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 554.612057] env[62460]: DEBUG nova.network.neutron [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 554.614887] env[62460]: DEBUG nova.network.neutron [-] [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 554.662821] env[62460]: DEBUG oslo_vmware.api [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]520967ff-ed71-6240-729d-f854b784a946, 'name': SearchDatastore_Task, 'duration_secs': 0.010694} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 554.664865] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 554.664865] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 554.664865] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 554.664865] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 554.665145] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 554.665145] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-78215540-b8fb-430f-a1a4-420b61f5c8db {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.686081] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 554.686512] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Folder [datastore1] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 554.688189] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2082d938-35f7-4865-ab1e-ad790a6861e6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.695542] env[62460]: DEBUG nova.policy [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '114c86641ab6422e9eafc6e27e0cc54e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2f7ef695e6084cd5837e00f0e7694fcd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 554.704655] env[62460]: DEBUG oslo_vmware.api [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Waiting for the task: (returnval){ [ 554.704655] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52ecabf3-691d-345e-d055-963698865f41" [ 554.704655] env[62460]: _type = "Task" [ 554.704655] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 554.717864] env[62460]: DEBUG oslo_vmware.api [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52ecabf3-691d-345e-d055-963698865f41, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 554.875467] env[62460]: INFO nova.scheduler.client.report [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] Deleted allocations for instance 7da5a822-6681-4dc6-b637-926ce425a042 [ 554.954331] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bce73fd-e31a-4d51-89cd-3aea6301f140 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.964501] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd20f1aa-029b-4cbc-a108-5d3a6364fcb1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.009950] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6afee5f6-d21d-4f66-805a-b588b335f55c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.020394] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55640e8d-c4ff-4ce9-a238-cb31acbc5247 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.037159] env[62460]: DEBUG nova.compute.provider_tree [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 555.113995] env[62460]: DEBUG nova.compute.manager [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 555.185892] env[62460]: DEBUG nova.network.neutron [req-25bfc2ce-e253-4a67-806f-272848d597af req-e74f4465-772e-4174-9608-9b1d9e38af8b service nova] [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 555.221023] env[62460]: DEBUG oslo_vmware.api [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52ecabf3-691d-345e-d055-963698865f41, 'name': SearchDatastore_Task, 'duration_secs': 0.026061} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 555.221792] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-74982b89-5f98-41e1-ac67-ab388f9a4066 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.228500] env[62460]: DEBUG oslo_vmware.api [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Waiting for the task: (returnval){ [ 555.228500] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]5228e7c1-57d7-a8f0-f393-c0b1087741ea" [ 555.228500] env[62460]: _type = "Task" [ 555.228500] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 555.240139] env[62460]: DEBUG oslo_vmware.api [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5228e7c1-57d7-a8f0-f393-c0b1087741ea, 'name': SearchDatastore_Task, 'duration_secs': 0.008641} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 555.240341] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 555.240877] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Copying Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 44a2ac83-3fff-4958-a1d3-b3884b42c211/44a2ac83-3fff-4958-a1d3-b3884b42c211.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 555.240985] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ab939b38-2170-46a6-a634-665a6e5633ea {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.250215] env[62460]: DEBUG oslo_vmware.api [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Waiting for the task: (returnval){ [ 555.250215] env[62460]: value = "task-1313418" [ 555.250215] env[62460]: _type = "Task" [ 555.250215] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 555.261444] env[62460]: DEBUG oslo_vmware.api [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Task: {'id': task-1313418, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 555.338867] env[62460]: DEBUG nova.network.neutron [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] Successfully created port: aa75f1cd-ce4c-485f-8cff-f4e7461216b6 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 555.383659] env[62460]: DEBUG oslo_concurrency.lockutils [None req-30878139-1902-4a27-a352-ac27783c9c40 tempest-ImagesOneServerTestJSON-1210704363 tempest-ImagesOneServerTestJSON-1210704363-project-member] Lock "7da5a822-6681-4dc6-b637-926ce425a042" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.708s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 555.463859] env[62460]: DEBUG nova.network.neutron [req-25bfc2ce-e253-4a67-806f-272848d597af req-e74f4465-772e-4174-9608-9b1d9e38af8b service nova] [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 555.506450] env[62460]: DEBUG oslo_concurrency.lockutils [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Acquiring lock "443e235d-32b2-4af7-bdae-3f30e8196ee3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 555.507065] env[62460]: DEBUG oslo_concurrency.lockutils [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Lock "443e235d-32b2-4af7-bdae-3f30e8196ee3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 555.543152] env[62460]: DEBUG nova.scheduler.client.report [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 555.764238] env[62460]: DEBUG oslo_vmware.api [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Task: {'id': task-1313418, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.488756} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 555.764755] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Copied Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 44a2ac83-3fff-4958-a1d3-b3884b42c211/44a2ac83-3fff-4958-a1d3-b3884b42c211.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 555.765129] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 555.766581] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8fc40e63-b9c9-42e9-9173-a2a8b5311488 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.773803] env[62460]: DEBUG oslo_vmware.api [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Waiting for the task: (returnval){ [ 555.773803] env[62460]: value = "task-1313419" [ 555.773803] env[62460]: _type = "Task" [ 555.773803] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 555.787399] env[62460]: DEBUG oslo_vmware.api [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Task: {'id': task-1313419, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 555.891628] env[62460]: DEBUG nova.compute.manager [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 555.968871] env[62460]: DEBUG oslo_concurrency.lockutils [req-25bfc2ce-e253-4a67-806f-272848d597af req-e74f4465-772e-4174-9608-9b1d9e38af8b service nova] Releasing lock "refresh_cache-ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 556.047818] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.952s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 556.050036] env[62460]: ERROR nova.compute.manager [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6d1179e6-551c-4986-b849-eceeefc23045, please check neutron logs for more information. [ 556.050036] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] Traceback (most recent call last): [ 556.050036] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 556.050036] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] self.driver.spawn(context, instance, image_meta, [ 556.050036] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 556.050036] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] self._vmops.spawn(context, instance, image_meta, injected_files, [ 556.050036] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 556.050036] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] vm_ref = self.build_virtual_machine(instance, [ 556.050036] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 556.050036] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] vif_infos = vmwarevif.get_vif_info(self._session, [ 556.050036] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 556.050462] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] for vif in network_info: [ 556.050462] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 556.050462] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] return self._sync_wrapper(fn, *args, **kwargs) [ 556.050462] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 556.050462] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] self.wait() [ 556.050462] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 556.050462] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] self[:] = self._gt.wait() [ 556.050462] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 556.050462] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] return self._exit_event.wait() [ 556.050462] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 556.050462] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] result = hub.switch() [ 556.050462] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 556.050462] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] return self.greenlet.switch() [ 556.050824] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 556.050824] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] result = function(*args, **kwargs) [ 556.050824] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 556.050824] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] return func(*args, **kwargs) [ 556.050824] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 556.050824] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] raise e [ 556.050824] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 556.050824] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] nwinfo = self.network_api.allocate_for_instance( [ 556.050824] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 556.050824] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] created_port_ids = self._update_ports_for_instance( [ 556.050824] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 556.050824] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] with excutils.save_and_reraise_exception(): [ 556.050824] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 556.051305] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] self.force_reraise() [ 556.051305] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 556.051305] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] raise self.value [ 556.051305] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 556.051305] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] updated_port = self._update_port( [ 556.051305] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 556.051305] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] _ensure_no_port_binding_failure(port) [ 556.051305] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 556.051305] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] raise exception.PortBindingFailed(port_id=port['id']) [ 556.051305] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] nova.exception.PortBindingFailed: Binding failed for port 6d1179e6-551c-4986-b849-eceeefc23045, please check neutron logs for more information. [ 556.051305] env[62460]: ERROR nova.compute.manager [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] [ 556.051656] env[62460]: DEBUG nova.compute.utils [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] Binding failed for port 6d1179e6-551c-4986-b849-eceeefc23045, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 556.051656] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.413s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 556.058019] env[62460]: DEBUG nova.compute.manager [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] Build of instance 073b92a2-8db9-4337-af2e-5e85c78d4b93 was re-scheduled: Binding failed for port 6d1179e6-551c-4986-b849-eceeefc23045, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 556.058019] env[62460]: DEBUG nova.compute.manager [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 556.058019] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Acquiring lock "refresh_cache-073b92a2-8db9-4337-af2e-5e85c78d4b93" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 556.058019] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Acquired lock "refresh_cache-073b92a2-8db9-4337-af2e-5e85c78d4b93" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 556.058019] env[62460]: DEBUG nova.network.neutron [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 556.125563] env[62460]: DEBUG nova.compute.manager [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 556.158637] env[62460]: DEBUG nova.virt.hardware [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 556.158852] env[62460]: DEBUG nova.virt.hardware [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 556.158954] env[62460]: DEBUG nova.virt.hardware [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 556.159211] env[62460]: DEBUG nova.virt.hardware [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 556.159400] env[62460]: DEBUG nova.virt.hardware [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 556.159874] env[62460]: DEBUG nova.virt.hardware [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 556.160166] env[62460]: DEBUG nova.virt.hardware [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 556.160568] env[62460]: DEBUG nova.virt.hardware [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 556.160776] env[62460]: DEBUG nova.virt.hardware [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 556.161084] env[62460]: DEBUG nova.virt.hardware [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 556.161263] env[62460]: DEBUG nova.virt.hardware [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 556.162632] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df89eeed-0fdc-496b-888e-25901061a48c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.173731] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bf3bb76-f12a-4672-b2c0-fc06ac6cb8f3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.289021] env[62460]: DEBUG oslo_vmware.api [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Task: {'id': task-1313419, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070412} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 556.289021] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 556.289700] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d68d3f01-847e-43e5-9062-fe6f0c7c8f76 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.316277] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Reconfiguring VM instance instance-00000001 to attach disk [datastore1] 44a2ac83-3fff-4958-a1d3-b3884b42c211/44a2ac83-3fff-4958-a1d3-b3884b42c211.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 556.316891] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e5d5a711-58e6-4597-a10e-346fbac056d6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.340050] env[62460]: DEBUG oslo_vmware.api [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Waiting for the task: (returnval){ [ 556.340050] env[62460]: value = "task-1313420" [ 556.340050] env[62460]: _type = "Task" [ 556.340050] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 556.348914] env[62460]: DEBUG oslo_vmware.api [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Task: {'id': task-1313420, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 556.419558] env[62460]: DEBUG nova.network.neutron [-] [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 556.424687] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 556.652838] env[62460]: DEBUG nova.network.neutron [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 556.852157] env[62460]: DEBUG oslo_vmware.api [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Task: {'id': task-1313420, 'name': ReconfigVM_Task, 'duration_secs': 0.29532} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 556.852625] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Reconfigured VM instance instance-00000001 to attach disk [datastore1] 44a2ac83-3fff-4958-a1d3-b3884b42c211/44a2ac83-3fff-4958-a1d3-b3884b42c211.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 556.854574] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-01b7452a-c319-4f9f-9938-3467367d325b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.857081] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d63b9b20-da5e-4eeb-a2e9-4ec722ea7719 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.866509] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f01e1e24-e9d9-4abb-9ee3-3f7ed9bc3853 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.870881] env[62460]: DEBUG oslo_vmware.api [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Waiting for the task: (returnval){ [ 556.870881] env[62460]: value = "task-1313421" [ 556.870881] env[62460]: _type = "Task" [ 556.870881] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 556.905064] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25e01db1-f56a-4375-97db-52e63742d22c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.907877] env[62460]: DEBUG oslo_vmware.api [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Task: {'id': task-1313421, 'name': Rename_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 556.913664] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9e1fcec-95a8-4f4c-8a98-8e53f9dcf708 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.929544] env[62460]: INFO nova.compute.manager [-] [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] Took 2.39 seconds to deallocate network for instance. [ 556.930148] env[62460]: DEBUG nova.compute.provider_tree [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 556.933669] env[62460]: DEBUG nova.compute.claims [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 556.933928] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 556.979242] env[62460]: DEBUG nova.network.neutron [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 557.039701] env[62460]: ERROR nova.compute.manager [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port aa75f1cd-ce4c-485f-8cff-f4e7461216b6, please check neutron logs for more information. [ 557.039701] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 557.039701] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 557.039701] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 557.039701] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 557.039701] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 557.039701] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 557.039701] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 557.039701] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 557.039701] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 557.039701] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 557.039701] env[62460]: ERROR nova.compute.manager raise self.value [ 557.039701] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 557.039701] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 557.039701] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 557.039701] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 557.040188] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 557.040188] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 557.040188] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port aa75f1cd-ce4c-485f-8cff-f4e7461216b6, please check neutron logs for more information. [ 557.040188] env[62460]: ERROR nova.compute.manager [ 557.040188] env[62460]: Traceback (most recent call last): [ 557.040188] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 557.040188] env[62460]: listener.cb(fileno) [ 557.040188] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 557.040188] env[62460]: result = function(*args, **kwargs) [ 557.040188] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 557.040188] env[62460]: return func(*args, **kwargs) [ 557.040188] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 557.040188] env[62460]: raise e [ 557.040188] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 557.040188] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 557.040188] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 557.040188] env[62460]: created_port_ids = self._update_ports_for_instance( [ 557.040188] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 557.040188] env[62460]: with excutils.save_and_reraise_exception(): [ 557.040188] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 557.040188] env[62460]: self.force_reraise() [ 557.040188] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 557.040188] env[62460]: raise self.value [ 557.040188] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 557.040188] env[62460]: updated_port = self._update_port( [ 557.040188] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 557.040188] env[62460]: _ensure_no_port_binding_failure(port) [ 557.040188] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 557.040188] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 557.041179] env[62460]: nova.exception.PortBindingFailed: Binding failed for port aa75f1cd-ce4c-485f-8cff-f4e7461216b6, please check neutron logs for more information. [ 557.041179] env[62460]: Removing descriptor: 18 [ 557.041179] env[62460]: ERROR nova.compute.manager [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port aa75f1cd-ce4c-485f-8cff-f4e7461216b6, please check neutron logs for more information. [ 557.041179] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] Traceback (most recent call last): [ 557.041179] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 557.041179] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] yield resources [ 557.041179] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 557.041179] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] self.driver.spawn(context, instance, image_meta, [ 557.041179] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 557.041179] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 557.041179] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 557.041179] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] vm_ref = self.build_virtual_machine(instance, [ 557.041613] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 557.041613] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] vif_infos = vmwarevif.get_vif_info(self._session, [ 557.041613] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 557.041613] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] for vif in network_info: [ 557.041613] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 557.041613] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] return self._sync_wrapper(fn, *args, **kwargs) [ 557.041613] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 557.041613] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] self.wait() [ 557.041613] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 557.041613] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] self[:] = self._gt.wait() [ 557.041613] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 557.041613] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] return self._exit_event.wait() [ 557.041613] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 557.041963] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] result = hub.switch() [ 557.041963] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 557.041963] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] return self.greenlet.switch() [ 557.041963] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 557.041963] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] result = function(*args, **kwargs) [ 557.041963] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 557.041963] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] return func(*args, **kwargs) [ 557.041963] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 557.041963] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] raise e [ 557.041963] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 557.041963] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] nwinfo = self.network_api.allocate_for_instance( [ 557.041963] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 557.041963] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] created_port_ids = self._update_ports_for_instance( [ 557.042327] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 557.042327] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] with excutils.save_and_reraise_exception(): [ 557.042327] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 557.042327] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] self.force_reraise() [ 557.042327] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 557.042327] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] raise self.value [ 557.042327] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 557.042327] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] updated_port = self._update_port( [ 557.042327] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 557.042327] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] _ensure_no_port_binding_failure(port) [ 557.042327] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 557.042327] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] raise exception.PortBindingFailed(port_id=port['id']) [ 557.042678] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] nova.exception.PortBindingFailed: Binding failed for port aa75f1cd-ce4c-485f-8cff-f4e7461216b6, please check neutron logs for more information. [ 557.042678] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] [ 557.042678] env[62460]: INFO nova.compute.manager [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] Terminating instance [ 557.044063] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] Acquiring lock "refresh_cache-d52580bc-7c11-47f0-a6c4-12941c48d75c" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 557.044063] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] Acquired lock "refresh_cache-d52580bc-7c11-47f0-a6c4-12941c48d75c" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 557.045852] env[62460]: DEBUG nova.network.neutron [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 557.368573] env[62460]: DEBUG nova.compute.manager [req-e24147e3-cc03-4e61-b442-2a0c96027805 req-4c6bc86e-53ee-4aec-bf8f-57efd16a5a0f service nova] [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] Received event network-vif-deleted-0de37ccf-02a2-4a86-9bb6-d49a72ab3f31 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 557.388465] env[62460]: DEBUG oslo_vmware.api [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Task: {'id': task-1313421, 'name': Rename_Task, 'duration_secs': 0.166468} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 557.389049] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 557.389049] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6d5118d5-05e5-4c67-83a6-9796ccfd0457 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.396293] env[62460]: DEBUG oslo_vmware.api [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Waiting for the task: (returnval){ [ 557.396293] env[62460]: value = "task-1313422" [ 557.396293] env[62460]: _type = "Task" [ 557.396293] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 557.406024] env[62460]: DEBUG oslo_vmware.api [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Task: {'id': task-1313422, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 557.434613] env[62460]: DEBUG nova.scheduler.client.report [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 557.482175] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Releasing lock "refresh_cache-073b92a2-8db9-4337-af2e-5e85c78d4b93" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 557.482457] env[62460]: DEBUG nova.compute.manager [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 557.482743] env[62460]: DEBUG nova.compute.manager [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 557.482958] env[62460]: DEBUG nova.network.neutron [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 557.513929] env[62460]: DEBUG nova.network.neutron [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 557.569153] env[62460]: DEBUG nova.network.neutron [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 557.702194] env[62460]: DEBUG nova.network.neutron [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 557.757545] env[62460]: DEBUG nova.compute.manager [req-bf38a90d-1f68-45cc-ae4e-8d3ba4f32135 req-64d1f0f7-a5d9-4676-8505-8ada0c467aeb service nova] [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] Received event network-changed-aa75f1cd-ce4c-485f-8cff-f4e7461216b6 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 557.757727] env[62460]: DEBUG nova.compute.manager [req-bf38a90d-1f68-45cc-ae4e-8d3ba4f32135 req-64d1f0f7-a5d9-4676-8505-8ada0c467aeb service nova] [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] Refreshing instance network info cache due to event network-changed-aa75f1cd-ce4c-485f-8cff-f4e7461216b6. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 557.757913] env[62460]: DEBUG oslo_concurrency.lockutils [req-bf38a90d-1f68-45cc-ae4e-8d3ba4f32135 req-64d1f0f7-a5d9-4676-8505-8ada0c467aeb service nova] Acquiring lock "refresh_cache-d52580bc-7c11-47f0-a6c4-12941c48d75c" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 557.908191] env[62460]: DEBUG oslo_vmware.api [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Task: {'id': task-1313422, 'name': PowerOnVM_Task, 'duration_secs': 0.469725} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 557.908570] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 557.909014] env[62460]: DEBUG nova.compute.manager [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 557.910039] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbdc5c70-c9fe-4625-8bd7-55301ed851d3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.942980] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.892s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 557.943663] env[62460]: ERROR nova.compute.manager [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 16f5912b-00e1-4f98-9484-4aabc70c360d, please check neutron logs for more information. [ 557.943663] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] Traceback (most recent call last): [ 557.943663] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 557.943663] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] self.driver.spawn(context, instance, image_meta, [ 557.943663] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 557.943663] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] self._vmops.spawn(context, instance, image_meta, injected_files, [ 557.943663] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 557.943663] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] vm_ref = self.build_virtual_machine(instance, [ 557.943663] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 557.943663] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] vif_infos = vmwarevif.get_vif_info(self._session, [ 557.943663] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 557.944063] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] for vif in network_info: [ 557.944063] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 557.944063] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] return self._sync_wrapper(fn, *args, **kwargs) [ 557.944063] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 557.944063] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] self.wait() [ 557.944063] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 557.944063] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] self[:] = self._gt.wait() [ 557.944063] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 557.944063] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] return self._exit_event.wait() [ 557.944063] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 557.944063] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] result = hub.switch() [ 557.944063] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 557.944063] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] return self.greenlet.switch() [ 557.944469] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 557.944469] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] result = function(*args, **kwargs) [ 557.944469] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 557.944469] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] return func(*args, **kwargs) [ 557.944469] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 557.944469] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] raise e [ 557.944469] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 557.944469] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] nwinfo = self.network_api.allocate_for_instance( [ 557.944469] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 557.944469] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] created_port_ids = self._update_ports_for_instance( [ 557.944469] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 557.944469] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] with excutils.save_and_reraise_exception(): [ 557.944469] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 557.944861] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] self.force_reraise() [ 557.944861] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 557.944861] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] raise self.value [ 557.944861] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 557.944861] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] updated_port = self._update_port( [ 557.944861] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 557.944861] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] _ensure_no_port_binding_failure(port) [ 557.944861] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 557.944861] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] raise exception.PortBindingFailed(port_id=port['id']) [ 557.944861] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] nova.exception.PortBindingFailed: Binding failed for port 16f5912b-00e1-4f98-9484-4aabc70c360d, please check neutron logs for more information. [ 557.944861] env[62460]: ERROR nova.compute.manager [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] [ 557.945213] env[62460]: DEBUG nova.compute.utils [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] Binding failed for port 16f5912b-00e1-4f98-9484-4aabc70c360d, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 557.945812] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.956s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.951099] env[62460]: DEBUG nova.compute.manager [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] Build of instance d16440e7-aa8a-4041-a4cd-d0ea84d8c027 was re-scheduled: Binding failed for port 16f5912b-00e1-4f98-9484-4aabc70c360d, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 557.951894] env[62460]: DEBUG nova.compute.manager [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 557.951894] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] Acquiring lock "refresh_cache-d16440e7-aa8a-4041-a4cd-d0ea84d8c027" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 557.951894] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] Acquired lock "refresh_cache-d16440e7-aa8a-4041-a4cd-d0ea84d8c027" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 557.952183] env[62460]: DEBUG nova.network.neutron [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 558.021505] env[62460]: DEBUG nova.network.neutron [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 558.206205] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] Releasing lock "refresh_cache-d52580bc-7c11-47f0-a6c4-12941c48d75c" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 558.208732] env[62460]: DEBUG nova.compute.manager [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 558.208732] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 558.208732] env[62460]: DEBUG oslo_concurrency.lockutils [req-bf38a90d-1f68-45cc-ae4e-8d3ba4f32135 req-64d1f0f7-a5d9-4676-8505-8ada0c467aeb service nova] Acquired lock "refresh_cache-d52580bc-7c11-47f0-a6c4-12941c48d75c" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 558.208732] env[62460]: DEBUG nova.network.neutron [req-bf38a90d-1f68-45cc-ae4e-8d3ba4f32135 req-64d1f0f7-a5d9-4676-8505-8ada0c467aeb service nova] [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] Refreshing network info cache for port aa75f1cd-ce4c-485f-8cff-f4e7461216b6 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 558.211581] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-512ab735-b1c7-474d-ae39-afc241c6123f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.222903] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2bfb8e5-c2a5-4062-b660-597762b5c4ae {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.254210] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d52580bc-7c11-47f0-a6c4-12941c48d75c could not be found. [ 558.254539] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 558.254731] env[62460]: INFO nova.compute.manager [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] Took 0.05 seconds to destroy the instance on the hypervisor. [ 558.254974] env[62460]: DEBUG oslo.service.loopingcall [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 558.255217] env[62460]: DEBUG nova.compute.manager [-] [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 558.255313] env[62460]: DEBUG nova.network.neutron [-] [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 558.284962] env[62460]: DEBUG nova.network.neutron [-] [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 558.430097] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 558.503531] env[62460]: DEBUG nova.network.neutron [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 558.524085] env[62460]: INFO nova.compute.manager [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 073b92a2-8db9-4337-af2e-5e85c78d4b93] Took 1.04 seconds to deallocate network for instance. [ 558.756244] env[62460]: DEBUG nova.network.neutron [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 558.763522] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b44ae037-ea0b-4c34-8aff-83e927c31d9b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.773143] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-156bfecd-3d36-4f46-8dcf-4e8f59ad70c4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.807191] env[62460]: DEBUG nova.network.neutron [-] [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 558.810092] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afe92982-2739-441a-a9d2-78d0645d6360 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.818982] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2f8820c-ab62-4799-930e-98a6f2cdabe6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.833662] env[62460]: DEBUG nova.compute.provider_tree [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 558.913267] env[62460]: DEBUG nova.network.neutron [req-bf38a90d-1f68-45cc-ae4e-8d3ba4f32135 req-64d1f0f7-a5d9-4676-8505-8ada0c467aeb service nova] [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 559.148936] env[62460]: DEBUG nova.network.neutron [req-bf38a90d-1f68-45cc-ae4e-8d3ba4f32135 req-64d1f0f7-a5d9-4676-8505-8ada0c467aeb service nova] [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 559.262196] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] Releasing lock "refresh_cache-d16440e7-aa8a-4041-a4cd-d0ea84d8c027" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 559.262384] env[62460]: DEBUG nova.compute.manager [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 559.262573] env[62460]: DEBUG nova.compute.manager [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 559.262743] env[62460]: DEBUG nova.network.neutron [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 559.298318] env[62460]: DEBUG nova.network.neutron [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 559.309946] env[62460]: INFO nova.compute.manager [-] [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] Took 1.05 seconds to deallocate network for instance. [ 559.314351] env[62460]: DEBUG nova.compute.claims [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 559.314688] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 559.337009] env[62460]: DEBUG nova.scheduler.client.report [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 559.564248] env[62460]: INFO nova.scheduler.client.report [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Deleted allocations for instance 073b92a2-8db9-4337-af2e-5e85c78d4b93 [ 559.652303] env[62460]: DEBUG oslo_concurrency.lockutils [req-bf38a90d-1f68-45cc-ae4e-8d3ba4f32135 req-64d1f0f7-a5d9-4676-8505-8ada0c467aeb service nova] Releasing lock "refresh_cache-d52580bc-7c11-47f0-a6c4-12941c48d75c" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 559.801781] env[62460]: DEBUG nova.network.neutron [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 559.843093] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.897s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 559.843742] env[62460]: ERROR nova.compute.manager [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f6eb304f-e7db-40f7-a462-0db7083694d9, please check neutron logs for more information. [ 559.843742] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] Traceback (most recent call last): [ 559.843742] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 559.843742] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] self.driver.spawn(context, instance, image_meta, [ 559.843742] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 559.843742] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] self._vmops.spawn(context, instance, image_meta, injected_files, [ 559.843742] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 559.843742] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] vm_ref = self.build_virtual_machine(instance, [ 559.843742] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 559.843742] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] vif_infos = vmwarevif.get_vif_info(self._session, [ 559.843742] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 559.844174] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] for vif in network_info: [ 559.844174] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 559.844174] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] return self._sync_wrapper(fn, *args, **kwargs) [ 559.844174] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 559.844174] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] self.wait() [ 559.844174] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 559.844174] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] self[:] = self._gt.wait() [ 559.844174] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 559.844174] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] return self._exit_event.wait() [ 559.844174] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 559.844174] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] result = hub.switch() [ 559.844174] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 559.844174] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] return self.greenlet.switch() [ 559.844597] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 559.844597] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] result = function(*args, **kwargs) [ 559.844597] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 559.844597] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] return func(*args, **kwargs) [ 559.844597] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 559.844597] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] raise e [ 559.844597] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 559.844597] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] nwinfo = self.network_api.allocate_for_instance( [ 559.844597] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 559.844597] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] created_port_ids = self._update_ports_for_instance( [ 559.844597] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 559.844597] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] with excutils.save_and_reraise_exception(): [ 559.844597] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 559.844966] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] self.force_reraise() [ 559.844966] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 559.844966] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] raise self.value [ 559.844966] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 559.844966] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] updated_port = self._update_port( [ 559.844966] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 559.844966] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] _ensure_no_port_binding_failure(port) [ 559.844966] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 559.844966] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] raise exception.PortBindingFailed(port_id=port['id']) [ 559.844966] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] nova.exception.PortBindingFailed: Binding failed for port f6eb304f-e7db-40f7-a462-0db7083694d9, please check neutron logs for more information. [ 559.844966] env[62460]: ERROR nova.compute.manager [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] [ 559.845295] env[62460]: DEBUG nova.compute.utils [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] Binding failed for port f6eb304f-e7db-40f7-a462-0db7083694d9, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 559.845954] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 14.095s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 559.846183] env[62460]: DEBUG nova.objects.instance [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62460) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 559.849236] env[62460]: DEBUG nova.compute.manager [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] Build of instance 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19 was re-scheduled: Binding failed for port f6eb304f-e7db-40f7-a462-0db7083694d9, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 559.851324] env[62460]: DEBUG nova.compute.manager [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 559.851324] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Acquiring lock "refresh_cache-48e50ab0-9d4e-4c5d-96d5-55399b3a8e19" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 559.851324] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Acquired lock "refresh_cache-48e50ab0-9d4e-4c5d-96d5-55399b3a8e19" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 559.851324] env[62460]: DEBUG nova.network.neutron [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 560.079267] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a59f54e9-fd5a-4489-9dfd-7c43f7b44462 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Lock "073b92a2-8db9-4337-af2e-5e85c78d4b93" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.088s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 560.311114] env[62460]: INFO nova.compute.manager [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] [instance: d16440e7-aa8a-4041-a4cd-d0ea84d8c027] Took 1.05 seconds to deallocate network for instance. [ 560.411035] env[62460]: DEBUG nova.network.neutron [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 560.562373] env[62460]: DEBUG nova.network.neutron [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 560.584938] env[62460]: DEBUG nova.compute.manager [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 560.615622] env[62460]: DEBUG nova.compute.manager [req-9babeac5-ef8a-4f7a-91aa-df30d6597d12 req-3fe4d83d-b725-452e-8f1b-8badf16f28aa service nova] [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] Received event network-vif-deleted-aa75f1cd-ce4c-485f-8cff-f4e7461216b6 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 560.657471] env[62460]: DEBUG oslo_concurrency.lockutils [None req-18fe540e-36dc-4e28-a3bd-6e40afcbc385 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Acquiring lock "44a2ac83-3fff-4958-a1d3-b3884b42c211" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.657764] env[62460]: DEBUG oslo_concurrency.lockutils [None req-18fe540e-36dc-4e28-a3bd-6e40afcbc385 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Lock "44a2ac83-3fff-4958-a1d3-b3884b42c211" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.657976] env[62460]: DEBUG oslo_concurrency.lockutils [None req-18fe540e-36dc-4e28-a3bd-6e40afcbc385 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Acquiring lock "44a2ac83-3fff-4958-a1d3-b3884b42c211-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.658203] env[62460]: DEBUG oslo_concurrency.lockutils [None req-18fe540e-36dc-4e28-a3bd-6e40afcbc385 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Lock "44a2ac83-3fff-4958-a1d3-b3884b42c211-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.658382] env[62460]: DEBUG oslo_concurrency.lockutils [None req-18fe540e-36dc-4e28-a3bd-6e40afcbc385 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Lock "44a2ac83-3fff-4958-a1d3-b3884b42c211-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 560.661337] env[62460]: INFO nova.compute.manager [None req-18fe540e-36dc-4e28-a3bd-6e40afcbc385 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Terminating instance [ 560.667362] env[62460]: DEBUG oslo_concurrency.lockutils [None req-18fe540e-36dc-4e28-a3bd-6e40afcbc385 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Acquiring lock "refresh_cache-44a2ac83-3fff-4958-a1d3-b3884b42c211" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 560.667362] env[62460]: DEBUG oslo_concurrency.lockutils [None req-18fe540e-36dc-4e28-a3bd-6e40afcbc385 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Acquired lock "refresh_cache-44a2ac83-3fff-4958-a1d3-b3884b42c211" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 560.667362] env[62460]: DEBUG nova.network.neutron [None req-18fe540e-36dc-4e28-a3bd-6e40afcbc385 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 560.861189] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5956fe0e-7960-4c86-be37-84cedb4b9c4d tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.015s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 560.863053] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.282s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.865826] env[62460]: INFO nova.compute.claims [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] [instance: 00544d2a-1a15-4347-abe4-3641b5bf0cfe] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 561.065367] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Releasing lock "refresh_cache-48e50ab0-9d4e-4c5d-96d5-55399b3a8e19" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 561.065367] env[62460]: DEBUG nova.compute.manager [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 561.065367] env[62460]: DEBUG nova.compute.manager [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 561.065607] env[62460]: DEBUG nova.network.neutron [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 561.092968] env[62460]: DEBUG nova.network.neutron [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 561.119818] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.187970] env[62460]: DEBUG nova.network.neutron [None req-18fe540e-36dc-4e28-a3bd-6e40afcbc385 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 561.270698] env[62460]: DEBUG nova.network.neutron [None req-18fe540e-36dc-4e28-a3bd-6e40afcbc385 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 561.384995] env[62460]: INFO nova.scheduler.client.report [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] Deleted allocations for instance d16440e7-aa8a-4041-a4cd-d0ea84d8c027 [ 561.596543] env[62460]: DEBUG nova.network.neutron [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 561.774806] env[62460]: DEBUG oslo_concurrency.lockutils [None req-18fe540e-36dc-4e28-a3bd-6e40afcbc385 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Releasing lock "refresh_cache-44a2ac83-3fff-4958-a1d3-b3884b42c211" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 561.775224] env[62460]: DEBUG nova.compute.manager [None req-18fe540e-36dc-4e28-a3bd-6e40afcbc385 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 561.775699] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-18fe540e-36dc-4e28-a3bd-6e40afcbc385 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 561.778866] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b5bda6d-1dd5-47cb-bc16-3d3ed5d1fed4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.788241] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-18fe540e-36dc-4e28-a3bd-6e40afcbc385 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 561.788241] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9225be87-795c-44fa-91a7-b3fba932853e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.794117] env[62460]: DEBUG oslo_vmware.api [None req-18fe540e-36dc-4e28-a3bd-6e40afcbc385 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Waiting for the task: (returnval){ [ 561.794117] env[62460]: value = "task-1313423" [ 561.794117] env[62460]: _type = "Task" [ 561.794117] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 561.808478] env[62460]: DEBUG oslo_vmware.api [None req-18fe540e-36dc-4e28-a3bd-6e40afcbc385 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': task-1313423, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 561.896423] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4793cf6a-9b1a-48e0-a3cb-1ab21d1756a2 tempest-ServersWithSpecificFlavorTestJSON-1149044980 tempest-ServersWithSpecificFlavorTestJSON-1149044980-project-member] Lock "d16440e7-aa8a-4041-a4cd-d0ea84d8c027" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.399s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 562.102248] env[62460]: INFO nova.compute.manager [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19] Took 1.04 seconds to deallocate network for instance. [ 562.169183] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-836f7e49-06db-459f-a52d-770581e37ad7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.178185] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2f92102-3bd7-440a-9007-b35a0d664847 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.215044] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c851a731-41cd-4c39-8ed2-ca0267c29e74 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.221040] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-736c0475-95a1-49a7-b333-8fb27b938030 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.236419] env[62460]: DEBUG nova.compute.provider_tree [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 562.307768] env[62460]: DEBUG oslo_vmware.api [None req-18fe540e-36dc-4e28-a3bd-6e40afcbc385 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': task-1313423, 'name': PowerOffVM_Task, 'duration_secs': 0.132564} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 562.307768] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-18fe540e-36dc-4e28-a3bd-6e40afcbc385 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 562.307768] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-18fe540e-36dc-4e28-a3bd-6e40afcbc385 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 562.307768] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9d5e1ae1-3f1a-407e-8c07-3e468518455f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.339693] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-18fe540e-36dc-4e28-a3bd-6e40afcbc385 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 562.340151] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-18fe540e-36dc-4e28-a3bd-6e40afcbc385 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Deleting contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 562.341545] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-18fe540e-36dc-4e28-a3bd-6e40afcbc385 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Deleting the datastore file [datastore1] 44a2ac83-3fff-4958-a1d3-b3884b42c211 {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 562.341545] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6f635879-534d-4e2d-b556-be2a466af900 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.347824] env[62460]: DEBUG oslo_vmware.api [None req-18fe540e-36dc-4e28-a3bd-6e40afcbc385 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Waiting for the task: (returnval){ [ 562.347824] env[62460]: value = "task-1313425" [ 562.347824] env[62460]: _type = "Task" [ 562.347824] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 562.357707] env[62460]: DEBUG oslo_vmware.api [None req-18fe540e-36dc-4e28-a3bd-6e40afcbc385 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': task-1313425, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 562.401374] env[62460]: DEBUG nova.compute.manager [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 562.739958] env[62460]: DEBUG nova.scheduler.client.report [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 562.860943] env[62460]: DEBUG oslo_vmware.api [None req-18fe540e-36dc-4e28-a3bd-6e40afcbc385 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Task: {'id': task-1313425, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.092484} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 562.861182] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-18fe540e-36dc-4e28-a3bd-6e40afcbc385 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 562.861421] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-18fe540e-36dc-4e28-a3bd-6e40afcbc385 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Deleted contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 562.861643] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-18fe540e-36dc-4e28-a3bd-6e40afcbc385 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 562.861844] env[62460]: INFO nova.compute.manager [None req-18fe540e-36dc-4e28-a3bd-6e40afcbc385 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Took 1.09 seconds to destroy the instance on the hypervisor. [ 562.862166] env[62460]: DEBUG oslo.service.loopingcall [None req-18fe540e-36dc-4e28-a3bd-6e40afcbc385 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 562.862539] env[62460]: DEBUG nova.compute.manager [-] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 562.862539] env[62460]: DEBUG nova.network.neutron [-] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 562.892021] env[62460]: DEBUG nova.network.neutron [-] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 562.930978] env[62460]: DEBUG oslo_concurrency.lockutils [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 563.144139] env[62460]: INFO nova.scheduler.client.report [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Deleted allocations for instance 48e50ab0-9d4e-4c5d-96d5-55399b3a8e19 [ 563.245313] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.383s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 563.248088] env[62460]: DEBUG nova.compute.manager [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] [instance: 00544d2a-1a15-4347-abe4-3641b5bf0cfe] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 563.250325] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.292s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.395424] env[62460]: DEBUG nova.network.neutron [-] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 563.654420] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6ad964b0-59a6-43cc-9199-75bf00806935 tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Lock "48e50ab0-9d4e-4c5d-96d5-55399b3a8e19" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.866s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 563.757497] env[62460]: DEBUG nova.compute.utils [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 563.764485] env[62460]: DEBUG nova.compute.manager [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] [instance: 00544d2a-1a15-4347-abe4-3641b5bf0cfe] Not allocating networking since 'none' was specified. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 563.863131] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Acquiring lock "3b71c366-cbd9-4b98-aa0d-c55b56e69231" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 563.863445] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Lock "3b71c366-cbd9-4b98-aa0d-c55b56e69231" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.901047] env[62460]: INFO nova.compute.manager [-] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Took 1.04 seconds to deallocate network for instance. [ 564.027670] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.028482] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 564.080018] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e435064a-0593-4fe8-9bf6-6b358a1e611b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.092942] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8efd9a5-8a1a-4951-a4eb-dd6ffa2aa1e0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.128021] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96a3a16b-a143-44ac-af07-df7f77571d18 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.137267] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccf7412b-03d9-4791-97f7-84fde7fc9262 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.153111] env[62460]: DEBUG nova.compute.provider_tree [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 564.161152] env[62460]: DEBUG nova.compute.manager [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 564.265313] env[62460]: DEBUG nova.compute.manager [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] [instance: 00544d2a-1a15-4347-abe4-3641b5bf0cfe] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 564.410176] env[62460]: DEBUG oslo_concurrency.lockutils [None req-18fe540e-36dc-4e28-a3bd-6e40afcbc385 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.444763] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Acquiring lock "01bbc9ec-f2ee-4a48-a33c-784861e81097" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.445091] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Lock "01bbc9ec-f2ee-4a48-a33c-784861e81097" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 564.656167] env[62460]: DEBUG nova.scheduler.client.report [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 564.695412] env[62460]: DEBUG oslo_concurrency.lockutils [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.160945] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.911s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 565.165017] env[62460]: ERROR nova.compute.manager [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 85fa9617-3127-47cb-8e24-46840450c4da, please check neutron logs for more information. [ 565.165017] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] Traceback (most recent call last): [ 565.165017] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 565.165017] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] self.driver.spawn(context, instance, image_meta, [ 565.165017] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 565.165017] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 565.165017] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 565.165017] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] vm_ref = self.build_virtual_machine(instance, [ 565.165017] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 565.165017] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] vif_infos = vmwarevif.get_vif_info(self._session, [ 565.165017] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 565.165441] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] for vif in network_info: [ 565.165441] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 565.165441] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] return self._sync_wrapper(fn, *args, **kwargs) [ 565.165441] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 565.165441] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] self.wait() [ 565.165441] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 565.165441] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] self[:] = self._gt.wait() [ 565.165441] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 565.165441] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] return self._exit_event.wait() [ 565.165441] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 565.165441] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] result = hub.switch() [ 565.165441] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 565.165441] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] return self.greenlet.switch() [ 565.165913] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 565.165913] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] result = function(*args, **kwargs) [ 565.165913] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 565.165913] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] return func(*args, **kwargs) [ 565.165913] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 565.165913] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] raise e [ 565.165913] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 565.165913] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] nwinfo = self.network_api.allocate_for_instance( [ 565.165913] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 565.165913] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] created_port_ids = self._update_ports_for_instance( [ 565.165913] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 565.165913] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] with excutils.save_and_reraise_exception(): [ 565.165913] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 565.166303] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] self.force_reraise() [ 565.166303] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 565.166303] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] raise self.value [ 565.166303] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 565.166303] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] updated_port = self._update_port( [ 565.166303] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 565.166303] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] _ensure_no_port_binding_failure(port) [ 565.166303] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 565.166303] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] raise exception.PortBindingFailed(port_id=port['id']) [ 565.166303] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] nova.exception.PortBindingFailed: Binding failed for port 85fa9617-3127-47cb-8e24-46840450c4da, please check neutron logs for more information. [ 565.166303] env[62460]: ERROR nova.compute.manager [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] [ 565.166702] env[62460]: DEBUG nova.compute.utils [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] Binding failed for port 85fa9617-3127-47cb-8e24-46840450c4da, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 565.166702] env[62460]: DEBUG oslo_concurrency.lockutils [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.874s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.166702] env[62460]: INFO nova.compute.claims [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 565.167629] env[62460]: DEBUG nova.compute.manager [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] Build of instance 72714a2e-a95d-495e-b163-f5168148fcdf was re-scheduled: Binding failed for port 85fa9617-3127-47cb-8e24-46840450c4da, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 565.168074] env[62460]: DEBUG nova.compute.manager [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 565.168297] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] Acquiring lock "refresh_cache-72714a2e-a95d-495e-b163-f5168148fcdf" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 565.168440] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] Acquired lock "refresh_cache-72714a2e-a95d-495e-b163-f5168148fcdf" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 565.168594] env[62460]: DEBUG nova.network.neutron [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 565.279640] env[62460]: DEBUG nova.compute.manager [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] [instance: 00544d2a-1a15-4347-abe4-3641b5bf0cfe] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 565.313934] env[62460]: DEBUG nova.virt.hardware [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 565.314259] env[62460]: DEBUG nova.virt.hardware [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 565.314464] env[62460]: DEBUG nova.virt.hardware [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 565.314713] env[62460]: DEBUG nova.virt.hardware [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 565.314895] env[62460]: DEBUG nova.virt.hardware [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 565.315133] env[62460]: DEBUG nova.virt.hardware [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 565.315436] env[62460]: DEBUG nova.virt.hardware [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 565.315556] env[62460]: DEBUG nova.virt.hardware [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 565.315761] env[62460]: DEBUG nova.virt.hardware [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 565.315981] env[62460]: DEBUG nova.virt.hardware [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 565.316229] env[62460]: DEBUG nova.virt.hardware [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 565.317202] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-906b3d85-d209-4d95-9f40-18ededbc2b25 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.326060] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca5c3f7f-7f3e-4dc1-90b6-06e88952e5ba {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.339766] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] [instance: 00544d2a-1a15-4347-abe4-3641b5bf0cfe] Instance VIF info [] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 565.345953] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Creating folder: Project (68a85ef19a7a4adea4d07018f5f54b44). Parent ref: group-v281134. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 565.346296] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5fb7733c-37ad-4f6c-9a14-1b9d8909181e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.356277] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Created folder: Project (68a85ef19a7a4adea4d07018f5f54b44) in parent group-v281134. [ 565.356549] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Creating folder: Instances. Parent ref: group-v281144. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 565.357010] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-39589156-99bb-4dd5-80c0-1032d58a8f16 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.365630] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Created folder: Instances in parent group-v281144. [ 565.365868] env[62460]: DEBUG oslo.service.loopingcall [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 565.366093] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 00544d2a-1a15-4347-abe4-3641b5bf0cfe] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 565.366319] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c1c3ffdf-5730-4961-a162-73af49f6834e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.385510] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 565.385510] env[62460]: value = "task-1313428" [ 565.385510] env[62460]: _type = "Task" [ 565.385510] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 565.393953] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313428, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 565.703873] env[62460]: DEBUG nova.network.neutron [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 565.899471] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313428, 'name': CreateVM_Task, 'duration_secs': 0.294905} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 565.906295] env[62460]: DEBUG nova.network.neutron [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 565.906295] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 00544d2a-1a15-4347-abe4-3641b5bf0cfe] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 565.906295] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 565.906295] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 565.906295] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 565.906532] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4f5f0f5b-993f-4dad-9baa-4edb721c405a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.911628] env[62460]: DEBUG oslo_vmware.api [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Waiting for the task: (returnval){ [ 565.911628] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]5200ea50-f19f-730a-e8b9-17950b4afebd" [ 565.911628] env[62460]: _type = "Task" [ 565.911628] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 565.925072] env[62460]: DEBUG oslo_vmware.api [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5200ea50-f19f-730a-e8b9-17950b4afebd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 566.409273] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] Releasing lock "refresh_cache-72714a2e-a95d-495e-b163-f5168148fcdf" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 566.409571] env[62460]: DEBUG nova.compute.manager [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 566.409753] env[62460]: DEBUG nova.compute.manager [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 566.409920] env[62460]: DEBUG nova.network.neutron [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 566.427396] env[62460]: DEBUG oslo_vmware.api [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5200ea50-f19f-730a-e8b9-17950b4afebd, 'name': SearchDatastore_Task, 'duration_secs': 0.010044} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 566.427756] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 566.428435] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] [instance: 00544d2a-1a15-4347-abe4-3641b5bf0cfe] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 566.428435] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 566.428435] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 566.428835] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 566.428835] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-59724e48-3f3d-4c0d-89b1-dc0a28717052 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.436716] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 566.436901] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 566.437604] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-19c11abf-5ba5-42c6-b43e-0bf6341ea743 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.448677] env[62460]: DEBUG oslo_vmware.api [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Waiting for the task: (returnval){ [ 566.448677] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52819cce-bbe0-6078-886e-ee240648142d" [ 566.448677] env[62460]: _type = "Task" [ 566.448677] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 566.454327] env[62460]: DEBUG nova.network.neutron [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 566.458857] env[62460]: DEBUG oslo_vmware.api [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52819cce-bbe0-6078-886e-ee240648142d, 'name': SearchDatastore_Task, 'duration_secs': 0.009003} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 566.463805] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b7210f5c-4782-4120-a8ae-e7511306e3a4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.469615] env[62460]: DEBUG oslo_vmware.api [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Waiting for the task: (returnval){ [ 566.469615] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]5246e4bb-cc10-a4b4-58e0-02a1638a5191" [ 566.469615] env[62460]: _type = "Task" [ 566.469615] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 566.483657] env[62460]: DEBUG oslo_vmware.api [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5246e4bb-cc10-a4b4-58e0-02a1638a5191, 'name': SearchDatastore_Task, 'duration_secs': 0.007918} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 566.484294] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 566.484294] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 00544d2a-1a15-4347-abe4-3641b5bf0cfe/00544d2a-1a15-4347-abe4-3641b5bf0cfe.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 566.484776] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f6600763-aad0-4b6c-8f73-ebd9cbac0dcf {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.491888] env[62460]: DEBUG oslo_vmware.api [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Waiting for the task: (returnval){ [ 566.491888] env[62460]: value = "task-1313429" [ 566.491888] env[62460]: _type = "Task" [ 566.491888] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 566.499200] env[62460]: DEBUG oslo_vmware.api [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Task: {'id': task-1313429, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 566.501888] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e67a6789-8735-496c-963e-4b79af66bffc {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.508502] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08e95630-fbb1-45eb-9668-cfaaf83f320a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.539931] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc0ebaa1-7e45-42d3-8307-031af228971a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.547985] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3af75d7-0256-4af6-9d57-7d5373277ecf {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.563170] env[62460]: DEBUG nova.compute.provider_tree [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 566.963260] env[62460]: DEBUG nova.network.neutron [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 567.009670] env[62460]: DEBUG oslo_vmware.api [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Task: {'id': task-1313429, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 567.068416] env[62460]: DEBUG nova.scheduler.client.report [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 567.475027] env[62460]: INFO nova.compute.manager [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] [instance: 72714a2e-a95d-495e-b163-f5168148fcdf] Took 1.06 seconds to deallocate network for instance. [ 567.503906] env[62460]: DEBUG oslo_vmware.api [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Task: {'id': task-1313429, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.557433} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 567.504635] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 00544d2a-1a15-4347-abe4-3641b5bf0cfe/00544d2a-1a15-4347-abe4-3641b5bf0cfe.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 567.505065] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] [instance: 00544d2a-1a15-4347-abe4-3641b5bf0cfe] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 567.506186] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-81a17ba3-f0c0-498b-b05e-779a3d5ad274 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.516976] env[62460]: DEBUG oslo_vmware.api [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Waiting for the task: (returnval){ [ 567.516976] env[62460]: value = "task-1313430" [ 567.516976] env[62460]: _type = "Task" [ 567.516976] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 567.524704] env[62460]: DEBUG oslo_vmware.api [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Task: {'id': task-1313430, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 567.573576] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] Acquiring lock "94703126-bb31-40af-b945-2ef5ff37e094" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 567.573910] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] Lock "94703126-bb31-40af-b945-2ef5ff37e094" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 567.577117] env[62460]: DEBUG oslo_concurrency.lockutils [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.413s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 567.577640] env[62460]: DEBUG nova.compute.manager [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 567.580332] env[62460]: DEBUG oslo_concurrency.lockutils [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.295s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 567.582179] env[62460]: INFO nova.compute.claims [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 568.029583] env[62460]: DEBUG oslo_vmware.api [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Task: {'id': task-1313430, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.31088} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 568.029583] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] [instance: 00544d2a-1a15-4347-abe4-3641b5bf0cfe] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 568.029712] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5f3db6f-b192-4909-983d-d6783f1a6c1a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.056796] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] [instance: 00544d2a-1a15-4347-abe4-3641b5bf0cfe] Reconfiguring VM instance instance-0000000c to attach disk [datastore1] 00544d2a-1a15-4347-abe4-3641b5bf0cfe/00544d2a-1a15-4347-abe4-3641b5bf0cfe.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 568.058080] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b755b493-4e33-4438-89cf-efce5ef1b481 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.083835] env[62460]: DEBUG oslo_vmware.api [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Waiting for the task: (returnval){ [ 568.083835] env[62460]: value = "task-1313431" [ 568.083835] env[62460]: _type = "Task" [ 568.083835] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 568.088638] env[62460]: DEBUG nova.compute.utils [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 568.098347] env[62460]: DEBUG nova.compute.manager [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 568.098347] env[62460]: DEBUG nova.network.neutron [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 568.103785] env[62460]: DEBUG oslo_vmware.api [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Task: {'id': task-1313431, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 568.256257] env[62460]: DEBUG nova.policy [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd1d1801c79014f89875d0da0bf54ec9e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '760a2bff3f6a4a4ab2c6d993b3334432', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 568.504989] env[62460]: DEBUG oslo_concurrency.lockutils [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] Acquiring lock "3426ace3-0cff-4119-85fe-31e681a8597b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 568.505772] env[62460]: DEBUG oslo_concurrency.lockutils [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] Lock "3426ace3-0cff-4119-85fe-31e681a8597b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 568.519034] env[62460]: INFO nova.scheduler.client.report [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] Deleted allocations for instance 72714a2e-a95d-495e-b163-f5168148fcdf [ 568.594687] env[62460]: DEBUG oslo_vmware.api [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Task: {'id': task-1313431, 'name': ReconfigVM_Task, 'duration_secs': 0.264418} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 568.595143] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] [instance: 00544d2a-1a15-4347-abe4-3641b5bf0cfe] Reconfigured VM instance instance-0000000c to attach disk [datastore1] 00544d2a-1a15-4347-abe4-3641b5bf0cfe/00544d2a-1a15-4347-abe4-3641b5bf0cfe.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 568.599514] env[62460]: DEBUG nova.compute.manager [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 568.603175] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-841d6c89-457e-44e9-80f8-1d0026901817 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.618190] env[62460]: DEBUG oslo_vmware.api [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Waiting for the task: (returnval){ [ 568.618190] env[62460]: value = "task-1313432" [ 568.618190] env[62460]: _type = "Task" [ 568.618190] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 568.632760] env[62460]: DEBUG oslo_vmware.api [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Task: {'id': task-1313432, 'name': Rename_Task} progress is 10%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 568.908591] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddd46884-d24c-4102-8493-d6ccb86bbe4a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.916352] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99ba40e8-8999-4228-bc5f-801721fe74ff {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.959959] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4641b8d1-26ba-4a50-bd61-ffbbf6c6ce32 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.967734] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fbbffa1-7e9c-4e14-9798-c0b2e3514a76 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.991173] env[62460]: DEBUG nova.compute.provider_tree [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 569.030224] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9bcf6ff0-24e4-41e0-81be-ef1da8cdac8b tempest-VolumesAssistedSnapshotsTest-370540121 tempest-VolumesAssistedSnapshotsTest-370540121-project-member] Lock "72714a2e-a95d-495e-b163-f5168148fcdf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.656s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 569.132665] env[62460]: DEBUG nova.network.neutron [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] Successfully created port: 357e567f-881f-453c-b117-bbf4d09e4a2f {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 569.134724] env[62460]: DEBUG oslo_vmware.api [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Task: {'id': task-1313432, 'name': Rename_Task, 'duration_secs': 0.133154} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 569.135156] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] [instance: 00544d2a-1a15-4347-abe4-3641b5bf0cfe] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 569.135612] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-68d8bbec-9640-45c5-98dd-625b3e1ac89c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.142596] env[62460]: DEBUG oslo_vmware.api [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Waiting for the task: (returnval){ [ 569.142596] env[62460]: value = "task-1313433" [ 569.142596] env[62460]: _type = "Task" [ 569.142596] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 569.151252] env[62460]: DEBUG oslo_vmware.api [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Task: {'id': task-1313433, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 569.260024] env[62460]: DEBUG oslo_concurrency.lockutils [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Acquiring lock "dba18c87-50a2-4ac8-8bd2-77f8f62b723c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 569.260024] env[62460]: DEBUG oslo_concurrency.lockutils [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Lock "dba18c87-50a2-4ac8-8bd2-77f8f62b723c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 569.496820] env[62460]: DEBUG nova.scheduler.client.report [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 569.534880] env[62460]: DEBUG nova.compute.manager [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 569.618021] env[62460]: DEBUG nova.compute.manager [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 569.663383] env[62460]: DEBUG oslo_vmware.api [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Task: {'id': task-1313433, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 569.671470] env[62460]: DEBUG nova.virt.hardware [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 569.671470] env[62460]: DEBUG nova.virt.hardware [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 569.671470] env[62460]: DEBUG nova.virt.hardware [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 569.671747] env[62460]: DEBUG nova.virt.hardware [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 569.671747] env[62460]: DEBUG nova.virt.hardware [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 569.671831] env[62460]: DEBUG nova.virt.hardware [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 569.672378] env[62460]: DEBUG nova.virt.hardware [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 569.672614] env[62460]: DEBUG nova.virt.hardware [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 569.673928] env[62460]: DEBUG nova.virt.hardware [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 569.673928] env[62460]: DEBUG nova.virt.hardware [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 569.673928] env[62460]: DEBUG nova.virt.hardware [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 569.674459] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f5c771b-2154-4d56-9680-3e4fc5fb8f79 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.685584] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dba44cd9-d0a8-4610-a36d-57944e7b8c30 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.006507] env[62460]: DEBUG oslo_concurrency.lockutils [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.426s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 570.007078] env[62460]: DEBUG nova.compute.manager [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 570.013257] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.587s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 570.017017] env[62460]: INFO nova.compute.claims [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 570.082574] env[62460]: DEBUG oslo_concurrency.lockutils [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 570.157994] env[62460]: DEBUG oslo_vmware.api [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Task: {'id': task-1313433, 'name': PowerOnVM_Task, 'duration_secs': 0.632878} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 570.157994] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] [instance: 00544d2a-1a15-4347-abe4-3641b5bf0cfe] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 570.157994] env[62460]: INFO nova.compute.manager [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] [instance: 00544d2a-1a15-4347-abe4-3641b5bf0cfe] Took 4.88 seconds to spawn the instance on the hypervisor. [ 570.157994] env[62460]: DEBUG nova.compute.manager [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] [instance: 00544d2a-1a15-4347-abe4-3641b5bf0cfe] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 570.161921] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e5a9ceb-6fce-4f62-86da-631ae86bbe34 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.519057] env[62460]: DEBUG nova.compute.utils [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 570.523792] env[62460]: DEBUG nova.compute.manager [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 570.523792] env[62460]: DEBUG nova.network.neutron [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 570.684661] env[62460]: INFO nova.compute.manager [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] [instance: 00544d2a-1a15-4347-abe4-3641b5bf0cfe] Took 22.13 seconds to build instance. [ 570.813423] env[62460]: DEBUG nova.policy [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd1d1801c79014f89875d0da0bf54ec9e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '760a2bff3f6a4a4ab2c6d993b3334432', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 571.027263] env[62460]: DEBUG nova.compute.manager [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 571.194748] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7520c52f-88e4-4946-b6a3-526c1e7c4ed1 tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Lock "00544d2a-1a15-4347-abe4-3641b5bf0cfe" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.870s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 571.382513] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-417c30e7-5bd8-480d-b167-da47b055e1f0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.392495] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbf2bf1c-e3a3-4581-8b0e-ffaaa518499f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.429995] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ee8ae0d-e99f-4b51-98e5-d0fc6777be2b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.438868] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0b4a2de-2c97-4c50-b77f-fdf645ba7bae {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.453763] env[62460]: DEBUG nova.compute.provider_tree [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 571.539608] env[62460]: DEBUG nova.compute.manager [None req-a9b1079b-196d-4eff-a867-06fb304938b4 tempest-ServerDiagnosticsV248Test-2008062123 tempest-ServerDiagnosticsV248Test-2008062123-project-admin] [instance: 00544d2a-1a15-4347-abe4-3641b5bf0cfe] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 571.540593] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02b4f87b-d8d9-4a64-a76f-ac270ca9a011 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.554259] env[62460]: INFO nova.compute.manager [None req-a9b1079b-196d-4eff-a867-06fb304938b4 tempest-ServerDiagnosticsV248Test-2008062123 tempest-ServerDiagnosticsV248Test-2008062123-project-admin] [instance: 00544d2a-1a15-4347-abe4-3641b5bf0cfe] Retrieving diagnostics [ 571.554820] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f1ae889-b3d1-4e56-a180-ddfb93664335 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.699996] env[62460]: DEBUG nova.compute.manager [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 571.957646] env[62460]: DEBUG nova.scheduler.client.report [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 572.051093] env[62460]: DEBUG nova.compute.manager [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 572.093136] env[62460]: DEBUG nova.virt.hardware [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 572.093531] env[62460]: DEBUG nova.virt.hardware [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 572.093749] env[62460]: DEBUG nova.virt.hardware [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 572.093990] env[62460]: DEBUG nova.virt.hardware [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 572.094233] env[62460]: DEBUG nova.virt.hardware [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 572.094695] env[62460]: DEBUG nova.virt.hardware [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 572.095016] env[62460]: DEBUG nova.virt.hardware [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 572.095355] env[62460]: DEBUG nova.virt.hardware [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 572.095635] env[62460]: DEBUG nova.virt.hardware [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 572.096236] env[62460]: DEBUG nova.virt.hardware [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 572.096548] env[62460]: DEBUG nova.virt.hardware [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 572.098736] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a11ba08-87a6-4e4a-8518-4963de030ea2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.112813] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c3424cc-9ede-4a74-864d-dcb8195c2b4d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.182659] env[62460]: DEBUG nova.network.neutron [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] Successfully created port: 53ec4989-7198-480a-ab0a-351544564777 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 572.228305] env[62460]: DEBUG oslo_concurrency.lockutils [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 572.464292] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.453s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 572.464842] env[62460]: DEBUG nova.compute.manager [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 572.468939] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.535s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 572.902580] env[62460]: DEBUG nova.compute.manager [req-35a90c99-ed2f-49ab-af99-a191771661fd req-9b51e45c-b11c-4f27-9d14-9973a0c3e792 service nova] [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] Received event network-changed-357e567f-881f-453c-b117-bbf4d09e4a2f {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 572.902580] env[62460]: DEBUG nova.compute.manager [req-35a90c99-ed2f-49ab-af99-a191771661fd req-9b51e45c-b11c-4f27-9d14-9973a0c3e792 service nova] [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] Refreshing instance network info cache due to event network-changed-357e567f-881f-453c-b117-bbf4d09e4a2f. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 572.902580] env[62460]: DEBUG oslo_concurrency.lockutils [req-35a90c99-ed2f-49ab-af99-a191771661fd req-9b51e45c-b11c-4f27-9d14-9973a0c3e792 service nova] Acquiring lock "refresh_cache-efef4327-22c8-4b59-9b26-89bcfe89bc59" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 572.902908] env[62460]: DEBUG oslo_concurrency.lockutils [req-35a90c99-ed2f-49ab-af99-a191771661fd req-9b51e45c-b11c-4f27-9d14-9973a0c3e792 service nova] Acquired lock "refresh_cache-efef4327-22c8-4b59-9b26-89bcfe89bc59" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 572.902908] env[62460]: DEBUG nova.network.neutron [req-35a90c99-ed2f-49ab-af99-a191771661fd req-9b51e45c-b11c-4f27-9d14-9973a0c3e792 service nova] [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] Refreshing network info cache for port 357e567f-881f-453c-b117-bbf4d09e4a2f {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 572.972919] env[62460]: DEBUG nova.compute.utils [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 572.973907] env[62460]: DEBUG nova.compute.manager [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 572.974381] env[62460]: DEBUG nova.network.neutron [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 572.984049] env[62460]: ERROR nova.compute.manager [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 357e567f-881f-453c-b117-bbf4d09e4a2f, please check neutron logs for more information. [ 572.984049] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 572.984049] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 572.984049] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 572.984049] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 572.984049] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 572.984049] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 572.984049] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 572.984049] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 572.984049] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 572.984049] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 572.984049] env[62460]: ERROR nova.compute.manager raise self.value [ 572.984049] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 572.984049] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 572.984049] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 572.984049] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 572.984639] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 572.984639] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 572.984639] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 357e567f-881f-453c-b117-bbf4d09e4a2f, please check neutron logs for more information. [ 572.984639] env[62460]: ERROR nova.compute.manager [ 572.984639] env[62460]: Traceback (most recent call last): [ 572.984639] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 572.984639] env[62460]: listener.cb(fileno) [ 572.984639] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 572.984639] env[62460]: result = function(*args, **kwargs) [ 572.984639] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 572.984639] env[62460]: return func(*args, **kwargs) [ 572.984639] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 572.984639] env[62460]: raise e [ 572.984639] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 572.984639] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 572.984639] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 572.984639] env[62460]: created_port_ids = self._update_ports_for_instance( [ 572.984639] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 572.984639] env[62460]: with excutils.save_and_reraise_exception(): [ 572.984639] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 572.984639] env[62460]: self.force_reraise() [ 572.984639] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 572.984639] env[62460]: raise self.value [ 572.984639] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 572.984639] env[62460]: updated_port = self._update_port( [ 572.984639] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 572.984639] env[62460]: _ensure_no_port_binding_failure(port) [ 572.984639] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 572.984639] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 572.985483] env[62460]: nova.exception.PortBindingFailed: Binding failed for port 357e567f-881f-453c-b117-bbf4d09e4a2f, please check neutron logs for more information. [ 572.985483] env[62460]: Removing descriptor: 17 [ 572.986788] env[62460]: ERROR nova.compute.manager [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 357e567f-881f-453c-b117-bbf4d09e4a2f, please check neutron logs for more information. [ 572.986788] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] Traceback (most recent call last): [ 572.986788] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 572.986788] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] yield resources [ 572.986788] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 572.986788] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] self.driver.spawn(context, instance, image_meta, [ 572.986788] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 572.986788] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] self._vmops.spawn(context, instance, image_meta, injected_files, [ 572.986788] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 572.986788] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] vm_ref = self.build_virtual_machine(instance, [ 572.986788] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 572.987174] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] vif_infos = vmwarevif.get_vif_info(self._session, [ 572.987174] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 572.987174] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] for vif in network_info: [ 572.987174] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 572.987174] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] return self._sync_wrapper(fn, *args, **kwargs) [ 572.987174] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 572.987174] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] self.wait() [ 572.987174] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 572.987174] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] self[:] = self._gt.wait() [ 572.987174] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 572.987174] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] return self._exit_event.wait() [ 572.987174] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 572.987174] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] result = hub.switch() [ 572.987544] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 572.987544] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] return self.greenlet.switch() [ 572.987544] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 572.987544] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] result = function(*args, **kwargs) [ 572.987544] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 572.987544] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] return func(*args, **kwargs) [ 572.987544] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 572.987544] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] raise e [ 572.987544] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 572.987544] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] nwinfo = self.network_api.allocate_for_instance( [ 572.987544] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 572.987544] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] created_port_ids = self._update_ports_for_instance( [ 572.987544] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 572.987957] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] with excutils.save_and_reraise_exception(): [ 572.987957] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 572.987957] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] self.force_reraise() [ 572.987957] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 572.987957] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] raise self.value [ 572.987957] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 572.987957] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] updated_port = self._update_port( [ 572.987957] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 572.987957] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] _ensure_no_port_binding_failure(port) [ 572.987957] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 572.987957] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] raise exception.PortBindingFailed(port_id=port['id']) [ 572.987957] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] nova.exception.PortBindingFailed: Binding failed for port 357e567f-881f-453c-b117-bbf4d09e4a2f, please check neutron logs for more information. [ 572.987957] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] [ 572.988332] env[62460]: INFO nova.compute.manager [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] Terminating instance [ 572.989562] env[62460]: DEBUG oslo_concurrency.lockutils [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Acquiring lock "refresh_cache-efef4327-22c8-4b59-9b26-89bcfe89bc59" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 573.072681] env[62460]: DEBUG nova.policy [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '499258f6f0bf4d62a727864e20eaad6d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8e2d7106a48e49419aa366c11915cdbb', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 573.344642] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c25deee-3b98-4885-b41d-8577a777bb7f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.353960] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef63356f-eb4f-43b7-a110-a0e65a61eeb4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.394904] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80f383e9-814e-4c2b-96aa-eef3520f6c79 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.404256] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe1b5ede-870d-4d80-91eb-3b723f30434d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.422480] env[62460]: DEBUG nova.compute.provider_tree [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 573.479911] env[62460]: DEBUG nova.compute.manager [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 573.485219] env[62460]: DEBUG nova.network.neutron [req-35a90c99-ed2f-49ab-af99-a191771661fd req-9b51e45c-b11c-4f27-9d14-9973a0c3e792 service nova] [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 573.530862] env[62460]: DEBUG oslo_concurrency.lockutils [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Acquiring lock "01a3e1be-d29b-45d7-987e-66a4395ae2a4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.531237] env[62460]: DEBUG oslo_concurrency.lockutils [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Lock "01a3e1be-d29b-45d7-987e-66a4395ae2a4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.731835] env[62460]: DEBUG nova.network.neutron [req-35a90c99-ed2f-49ab-af99-a191771661fd req-9b51e45c-b11c-4f27-9d14-9973a0c3e792 service nova] [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 573.926789] env[62460]: DEBUG nova.scheduler.client.report [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 574.233965] env[62460]: DEBUG oslo_concurrency.lockutils [req-35a90c99-ed2f-49ab-af99-a191771661fd req-9b51e45c-b11c-4f27-9d14-9973a0c3e792 service nova] Releasing lock "refresh_cache-efef4327-22c8-4b59-9b26-89bcfe89bc59" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 574.234535] env[62460]: DEBUG oslo_concurrency.lockutils [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Acquired lock "refresh_cache-efef4327-22c8-4b59-9b26-89bcfe89bc59" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 574.234639] env[62460]: DEBUG nova.network.neutron [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 574.256132] env[62460]: DEBUG nova.network.neutron [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] Successfully created port: 0978e846-4023-4bcd-867b-55313312ab35 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 574.435401] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.966s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 574.436140] env[62460]: ERROR nova.compute.manager [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0de37ccf-02a2-4a86-9bb6-d49a72ab3f31, please check neutron logs for more information. [ 574.436140] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] Traceback (most recent call last): [ 574.436140] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 574.436140] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] self.driver.spawn(context, instance, image_meta, [ 574.436140] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 574.436140] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 574.436140] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 574.436140] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] vm_ref = self.build_virtual_machine(instance, [ 574.436140] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 574.436140] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] vif_infos = vmwarevif.get_vif_info(self._session, [ 574.436140] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 574.436569] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] for vif in network_info: [ 574.436569] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 574.436569] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] return self._sync_wrapper(fn, *args, **kwargs) [ 574.436569] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 574.436569] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] self.wait() [ 574.436569] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 574.436569] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] self[:] = self._gt.wait() [ 574.436569] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 574.436569] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] return self._exit_event.wait() [ 574.436569] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 574.436569] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] result = hub.switch() [ 574.436569] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 574.436569] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] return self.greenlet.switch() [ 574.436930] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 574.436930] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] result = function(*args, **kwargs) [ 574.436930] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 574.436930] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] return func(*args, **kwargs) [ 574.436930] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 574.436930] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] raise e [ 574.436930] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 574.436930] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] nwinfo = self.network_api.allocate_for_instance( [ 574.436930] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 574.436930] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] created_port_ids = self._update_ports_for_instance( [ 574.436930] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 574.436930] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] with excutils.save_and_reraise_exception(): [ 574.436930] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 574.437524] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] self.force_reraise() [ 574.437524] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 574.437524] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] raise self.value [ 574.437524] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 574.437524] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] updated_port = self._update_port( [ 574.437524] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 574.437524] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] _ensure_no_port_binding_failure(port) [ 574.437524] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 574.437524] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] raise exception.PortBindingFailed(port_id=port['id']) [ 574.437524] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] nova.exception.PortBindingFailed: Binding failed for port 0de37ccf-02a2-4a86-9bb6-d49a72ab3f31, please check neutron logs for more information. [ 574.437524] env[62460]: ERROR nova.compute.manager [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] [ 574.437833] env[62460]: DEBUG nova.compute.utils [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] Binding failed for port 0de37ccf-02a2-4a86-9bb6-d49a72ab3f31, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 574.438364] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 16.008s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 574.439854] env[62460]: DEBUG nova.objects.instance [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62460) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 574.442937] env[62460]: DEBUG nova.compute.manager [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] Build of instance ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3 was re-scheduled: Binding failed for port 0de37ccf-02a2-4a86-9bb6-d49a72ab3f31, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 574.447014] env[62460]: DEBUG nova.compute.manager [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 574.447014] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Acquiring lock "refresh_cache-ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 574.447014] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Acquired lock "refresh_cache-ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 574.447014] env[62460]: DEBUG nova.network.neutron [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 574.498998] env[62460]: DEBUG nova.compute.manager [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 574.530241] env[62460]: DEBUG nova.virt.hardware [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 574.530493] env[62460]: DEBUG nova.virt.hardware [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 574.530708] env[62460]: DEBUG nova.virt.hardware [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 574.530848] env[62460]: DEBUG nova.virt.hardware [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 574.530962] env[62460]: DEBUG nova.virt.hardware [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 574.531505] env[62460]: DEBUG nova.virt.hardware [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 574.531876] env[62460]: DEBUG nova.virt.hardware [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 574.532040] env[62460]: DEBUG nova.virt.hardware [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 574.532122] env[62460]: DEBUG nova.virt.hardware [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 574.532518] env[62460]: DEBUG nova.virt.hardware [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 574.532518] env[62460]: DEBUG nova.virt.hardware [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 574.533371] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-335ca4bb-5b66-4c75-9369-9041e7ab27b9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.543856] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cf0d8f9-7219-492a-9ab2-10d61d996a2e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.785604] env[62460]: DEBUG nova.network.neutron [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 575.007239] env[62460]: DEBUG nova.network.neutron [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 575.083038] env[62460]: DEBUG nova.network.neutron [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 575.298619] env[62460]: DEBUG nova.compute.manager [req-68731e3b-f183-4480-82ac-5eacd5955458 req-b1f6f2f0-a657-41db-b1fd-149589d47575 service nova] [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] Received event network-vif-deleted-357e567f-881f-453c-b117-bbf4d09e4a2f {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 575.453865] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6c08d1c5-1e12-4bc7-9933-be951be1c13f tempest-ServersAdmin275Test-448279504 tempest-ServersAdmin275Test-448279504-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.015s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 575.454754] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.140s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 575.516569] env[62460]: DEBUG oslo_concurrency.lockutils [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Releasing lock "refresh_cache-efef4327-22c8-4b59-9b26-89bcfe89bc59" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 575.516569] env[62460]: DEBUG nova.compute.manager [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 575.516569] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 575.516569] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-503bbfa5-6ff2-4fa7-8c86-0ae62dd24004 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.528309] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b22a770-eb40-432c-a0f5-f01ec442f0de {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.541324] env[62460]: DEBUG nova.network.neutron [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 575.552334] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance efef4327-22c8-4b59-9b26-89bcfe89bc59 could not be found. [ 575.552720] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 575.553016] env[62460]: INFO nova.compute.manager [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] Took 0.04 seconds to destroy the instance on the hypervisor. [ 575.553307] env[62460]: DEBUG oslo.service.loopingcall [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 575.554032] env[62460]: DEBUG nova.compute.manager [-] [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 575.554137] env[62460]: DEBUG nova.network.neutron [-] [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 575.618698] env[62460]: DEBUG nova.network.neutron [-] [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 576.043940] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Releasing lock "refresh_cache-ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 576.044227] env[62460]: DEBUG nova.compute.manager [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 576.044661] env[62460]: DEBUG nova.compute.manager [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 576.044661] env[62460]: DEBUG nova.network.neutron [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 576.121886] env[62460]: DEBUG nova.network.neutron [-] [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 576.125561] env[62460]: DEBUG nova.network.neutron [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 576.288626] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41bb5fce-fb28-466a-afdf-cdf3781e340f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.298532] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe83df86-e843-4f5a-8704-018c1bacf1bd {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.335538] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89cabff7-135d-4a0c-9cfd-3bcd9efcd954 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.344642] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0771edbd-7a2b-4012-b1ab-a11aa99d0270 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.358775] env[62460]: DEBUG nova.compute.provider_tree [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 576.631505] env[62460]: INFO nova.compute.manager [-] [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] Took 1.07 seconds to deallocate network for instance. [ 576.631505] env[62460]: DEBUG nova.network.neutron [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 576.635025] env[62460]: DEBUG nova.compute.claims [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 576.635533] env[62460]: DEBUG oslo_concurrency.lockutils [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 576.865163] env[62460]: DEBUG nova.scheduler.client.report [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 577.135038] env[62460]: INFO nova.compute.manager [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3] Took 1.09 seconds to deallocate network for instance. [ 577.368376] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.913s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 577.369366] env[62460]: ERROR nova.compute.manager [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port aa75f1cd-ce4c-485f-8cff-f4e7461216b6, please check neutron logs for more information. [ 577.369366] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] Traceback (most recent call last): [ 577.369366] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 577.369366] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] self.driver.spawn(context, instance, image_meta, [ 577.369366] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 577.369366] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 577.369366] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 577.369366] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] vm_ref = self.build_virtual_machine(instance, [ 577.369366] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 577.369366] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] vif_infos = vmwarevif.get_vif_info(self._session, [ 577.369366] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 577.369877] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] for vif in network_info: [ 577.369877] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 577.369877] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] return self._sync_wrapper(fn, *args, **kwargs) [ 577.369877] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 577.369877] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] self.wait() [ 577.369877] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 577.369877] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] self[:] = self._gt.wait() [ 577.369877] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 577.369877] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] return self._exit_event.wait() [ 577.369877] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 577.369877] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] result = hub.switch() [ 577.369877] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 577.369877] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] return self.greenlet.switch() [ 577.370380] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 577.370380] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] result = function(*args, **kwargs) [ 577.370380] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 577.370380] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] return func(*args, **kwargs) [ 577.370380] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 577.370380] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] raise e [ 577.370380] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 577.370380] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] nwinfo = self.network_api.allocate_for_instance( [ 577.370380] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 577.370380] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] created_port_ids = self._update_ports_for_instance( [ 577.370380] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 577.370380] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] with excutils.save_and_reraise_exception(): [ 577.370380] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 577.370801] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] self.force_reraise() [ 577.370801] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 577.370801] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] raise self.value [ 577.370801] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 577.370801] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] updated_port = self._update_port( [ 577.370801] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 577.370801] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] _ensure_no_port_binding_failure(port) [ 577.370801] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 577.370801] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] raise exception.PortBindingFailed(port_id=port['id']) [ 577.370801] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] nova.exception.PortBindingFailed: Binding failed for port aa75f1cd-ce4c-485f-8cff-f4e7461216b6, please check neutron logs for more information. [ 577.370801] env[62460]: ERROR nova.compute.manager [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] [ 577.371195] env[62460]: DEBUG nova.compute.utils [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] Binding failed for port aa75f1cd-ce4c-485f-8cff-f4e7461216b6, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 577.372582] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.252s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 577.374220] env[62460]: INFO nova.compute.claims [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 577.380759] env[62460]: DEBUG nova.compute.manager [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] Build of instance d52580bc-7c11-47f0-a6c4-12941c48d75c was re-scheduled: Binding failed for port aa75f1cd-ce4c-485f-8cff-f4e7461216b6, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 577.381268] env[62460]: DEBUG nova.compute.manager [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 577.381502] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] Acquiring lock "refresh_cache-d52580bc-7c11-47f0-a6c4-12941c48d75c" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 577.381654] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] Acquired lock "refresh_cache-d52580bc-7c11-47f0-a6c4-12941c48d75c" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 577.381818] env[62460]: DEBUG nova.network.neutron [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 577.516750] env[62460]: DEBUG nova.compute.manager [req-ba2e6a99-58ee-4714-8864-6b837a99a8bd req-508a1ce4-d973-4e01-9b47-5c207cd7f3f9 service nova] [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] Received event network-changed-0978e846-4023-4bcd-867b-55313312ab35 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 577.517011] env[62460]: DEBUG nova.compute.manager [req-ba2e6a99-58ee-4714-8864-6b837a99a8bd req-508a1ce4-d973-4e01-9b47-5c207cd7f3f9 service nova] [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] Refreshing instance network info cache due to event network-changed-0978e846-4023-4bcd-867b-55313312ab35. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 577.517172] env[62460]: DEBUG oslo_concurrency.lockutils [req-ba2e6a99-58ee-4714-8864-6b837a99a8bd req-508a1ce4-d973-4e01-9b47-5c207cd7f3f9 service nova] Acquiring lock "refresh_cache-a3f7edbe-4bba-4fff-9e62-99b7a85f971b" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 577.517313] env[62460]: DEBUG oslo_concurrency.lockutils [req-ba2e6a99-58ee-4714-8864-6b837a99a8bd req-508a1ce4-d973-4e01-9b47-5c207cd7f3f9 service nova] Acquired lock "refresh_cache-a3f7edbe-4bba-4fff-9e62-99b7a85f971b" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 577.517616] env[62460]: DEBUG nova.network.neutron [req-ba2e6a99-58ee-4714-8864-6b837a99a8bd req-508a1ce4-d973-4e01-9b47-5c207cd7f3f9 service nova] [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] Refreshing network info cache for port 0978e846-4023-4bcd-867b-55313312ab35 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 577.544519] env[62460]: DEBUG nova.compute.manager [req-e2ba2685-590d-42f9-a848-894da55787c9 req-7a712bcd-63ec-4499-a5b8-e3c6a07aa609 service nova] [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] Received event network-changed-53ec4989-7198-480a-ab0a-351544564777 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 577.544759] env[62460]: DEBUG nova.compute.manager [req-e2ba2685-590d-42f9-a848-894da55787c9 req-7a712bcd-63ec-4499-a5b8-e3c6a07aa609 service nova] [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] Refreshing instance network info cache due to event network-changed-53ec4989-7198-480a-ab0a-351544564777. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 577.544913] env[62460]: DEBUG oslo_concurrency.lockutils [req-e2ba2685-590d-42f9-a848-894da55787c9 req-7a712bcd-63ec-4499-a5b8-e3c6a07aa609 service nova] Acquiring lock "refresh_cache-8a2699f3-3cc4-4ec3-899a-216e09ad086a" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 577.553524] env[62460]: DEBUG oslo_concurrency.lockutils [req-e2ba2685-590d-42f9-a848-894da55787c9 req-7a712bcd-63ec-4499-a5b8-e3c6a07aa609 service nova] Acquired lock "refresh_cache-8a2699f3-3cc4-4ec3-899a-216e09ad086a" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 577.553524] env[62460]: DEBUG nova.network.neutron [req-e2ba2685-590d-42f9-a848-894da55787c9 req-7a712bcd-63ec-4499-a5b8-e3c6a07aa609 service nova] [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] Refreshing network info cache for port 53ec4989-7198-480a-ab0a-351544564777 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 577.708414] env[62460]: ERROR nova.compute.manager [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0978e846-4023-4bcd-867b-55313312ab35, please check neutron logs for more information. [ 577.708414] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 577.708414] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 577.708414] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 577.708414] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 577.708414] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 577.708414] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 577.708414] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 577.708414] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 577.708414] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 577.708414] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 577.708414] env[62460]: ERROR nova.compute.manager raise self.value [ 577.708414] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 577.708414] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 577.708414] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 577.708414] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 577.709235] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 577.709235] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 577.709235] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0978e846-4023-4bcd-867b-55313312ab35, please check neutron logs for more information. [ 577.709235] env[62460]: ERROR nova.compute.manager [ 577.709235] env[62460]: Traceback (most recent call last): [ 577.709235] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 577.709235] env[62460]: listener.cb(fileno) [ 577.709235] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 577.709235] env[62460]: result = function(*args, **kwargs) [ 577.709235] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 577.709235] env[62460]: return func(*args, **kwargs) [ 577.709235] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 577.709235] env[62460]: raise e [ 577.709235] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 577.709235] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 577.709235] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 577.709235] env[62460]: created_port_ids = self._update_ports_for_instance( [ 577.709235] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 577.709235] env[62460]: with excutils.save_and_reraise_exception(): [ 577.709235] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 577.709235] env[62460]: self.force_reraise() [ 577.709235] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 577.709235] env[62460]: raise self.value [ 577.709235] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 577.709235] env[62460]: updated_port = self._update_port( [ 577.709235] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 577.709235] env[62460]: _ensure_no_port_binding_failure(port) [ 577.709235] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 577.709235] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 577.710643] env[62460]: nova.exception.PortBindingFailed: Binding failed for port 0978e846-4023-4bcd-867b-55313312ab35, please check neutron logs for more information. [ 577.710643] env[62460]: Removing descriptor: 19 [ 577.710643] env[62460]: ERROR nova.compute.manager [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0978e846-4023-4bcd-867b-55313312ab35, please check neutron logs for more information. [ 577.710643] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] Traceback (most recent call last): [ 577.710643] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 577.710643] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] yield resources [ 577.710643] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 577.710643] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] self.driver.spawn(context, instance, image_meta, [ 577.710643] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 577.710643] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 577.710643] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 577.710643] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] vm_ref = self.build_virtual_machine(instance, [ 577.711716] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 577.711716] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] vif_infos = vmwarevif.get_vif_info(self._session, [ 577.711716] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 577.711716] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] for vif in network_info: [ 577.711716] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 577.711716] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] return self._sync_wrapper(fn, *args, **kwargs) [ 577.711716] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 577.711716] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] self.wait() [ 577.711716] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 577.711716] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] self[:] = self._gt.wait() [ 577.711716] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 577.711716] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] return self._exit_event.wait() [ 577.711716] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 577.714843] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] result = hub.switch() [ 577.714843] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 577.714843] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] return self.greenlet.switch() [ 577.714843] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 577.714843] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] result = function(*args, **kwargs) [ 577.714843] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 577.714843] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] return func(*args, **kwargs) [ 577.714843] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 577.714843] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] raise e [ 577.714843] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 577.714843] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] nwinfo = self.network_api.allocate_for_instance( [ 577.714843] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 577.714843] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] created_port_ids = self._update_ports_for_instance( [ 577.715327] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 577.715327] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] with excutils.save_and_reraise_exception(): [ 577.715327] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 577.715327] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] self.force_reraise() [ 577.715327] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 577.715327] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] raise self.value [ 577.715327] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 577.715327] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] updated_port = self._update_port( [ 577.715327] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 577.715327] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] _ensure_no_port_binding_failure(port) [ 577.715327] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 577.715327] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] raise exception.PortBindingFailed(port_id=port['id']) [ 577.715691] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] nova.exception.PortBindingFailed: Binding failed for port 0978e846-4023-4bcd-867b-55313312ab35, please check neutron logs for more information. [ 577.715691] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] [ 577.715691] env[62460]: INFO nova.compute.manager [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] Terminating instance [ 577.716644] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] Acquiring lock "refresh_cache-a3f7edbe-4bba-4fff-9e62-99b7a85f971b" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 577.926926] env[62460]: DEBUG nova.network.neutron [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 578.079797] env[62460]: DEBUG nova.network.neutron [req-ba2e6a99-58ee-4714-8864-6b837a99a8bd req-508a1ce4-d973-4e01-9b47-5c207cd7f3f9 service nova] [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 578.173095] env[62460]: INFO nova.scheduler.client.report [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Deleted allocations for instance ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3 [ 578.403114] env[62460]: DEBUG nova.network.neutron [req-ba2e6a99-58ee-4714-8864-6b837a99a8bd req-508a1ce4-d973-4e01-9b47-5c207cd7f3f9 service nova] [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 578.404503] env[62460]: DEBUG nova.network.neutron [req-e2ba2685-590d-42f9-a848-894da55787c9 req-7a712bcd-63ec-4499-a5b8-e3c6a07aa609 service nova] [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 578.494277] env[62460]: DEBUG nova.network.neutron [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 578.532197] env[62460]: ERROR nova.compute.manager [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 53ec4989-7198-480a-ab0a-351544564777, please check neutron logs for more information. [ 578.532197] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 578.532197] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 578.532197] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 578.532197] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 578.532197] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 578.532197] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 578.532197] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 578.532197] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 578.532197] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 578.532197] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 578.532197] env[62460]: ERROR nova.compute.manager raise self.value [ 578.532197] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 578.532197] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 578.532197] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 578.532197] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 578.533133] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 578.533133] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 578.533133] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 53ec4989-7198-480a-ab0a-351544564777, please check neutron logs for more information. [ 578.533133] env[62460]: ERROR nova.compute.manager [ 578.533133] env[62460]: Traceback (most recent call last): [ 578.533133] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 578.533133] env[62460]: listener.cb(fileno) [ 578.533133] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 578.533133] env[62460]: result = function(*args, **kwargs) [ 578.533133] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 578.533133] env[62460]: return func(*args, **kwargs) [ 578.533133] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 578.533133] env[62460]: raise e [ 578.533133] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 578.533133] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 578.533133] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 578.533133] env[62460]: created_port_ids = self._update_ports_for_instance( [ 578.533133] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 578.533133] env[62460]: with excutils.save_and_reraise_exception(): [ 578.533133] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 578.533133] env[62460]: self.force_reraise() [ 578.533133] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 578.533133] env[62460]: raise self.value [ 578.533133] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 578.533133] env[62460]: updated_port = self._update_port( [ 578.533133] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 578.533133] env[62460]: _ensure_no_port_binding_failure(port) [ 578.533133] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 578.533133] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 578.534031] env[62460]: nova.exception.PortBindingFailed: Binding failed for port 53ec4989-7198-480a-ab0a-351544564777, please check neutron logs for more information. [ 578.534031] env[62460]: Removing descriptor: 18 [ 578.534031] env[62460]: ERROR nova.compute.manager [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 53ec4989-7198-480a-ab0a-351544564777, please check neutron logs for more information. [ 578.534031] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] Traceback (most recent call last): [ 578.534031] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 578.534031] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] yield resources [ 578.534031] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 578.534031] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] self.driver.spawn(context, instance, image_meta, [ 578.534031] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 578.534031] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 578.534031] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 578.534031] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] vm_ref = self.build_virtual_machine(instance, [ 578.534459] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 578.534459] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] vif_infos = vmwarevif.get_vif_info(self._session, [ 578.534459] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 578.534459] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] for vif in network_info: [ 578.534459] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 578.534459] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] return self._sync_wrapper(fn, *args, **kwargs) [ 578.534459] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 578.534459] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] self.wait() [ 578.534459] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 578.534459] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] self[:] = self._gt.wait() [ 578.534459] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 578.534459] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] return self._exit_event.wait() [ 578.534459] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 578.534839] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] result = hub.switch() [ 578.534839] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 578.534839] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] return self.greenlet.switch() [ 578.534839] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 578.534839] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] result = function(*args, **kwargs) [ 578.534839] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 578.534839] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] return func(*args, **kwargs) [ 578.534839] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 578.534839] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] raise e [ 578.534839] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 578.534839] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] nwinfo = self.network_api.allocate_for_instance( [ 578.534839] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 578.534839] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] created_port_ids = self._update_ports_for_instance( [ 578.536334] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 578.536334] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] with excutils.save_and_reraise_exception(): [ 578.536334] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 578.536334] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] self.force_reraise() [ 578.536334] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 578.536334] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] raise self.value [ 578.536334] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 578.536334] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] updated_port = self._update_port( [ 578.536334] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 578.536334] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] _ensure_no_port_binding_failure(port) [ 578.536334] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 578.536334] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] raise exception.PortBindingFailed(port_id=port['id']) [ 578.536853] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] nova.exception.PortBindingFailed: Binding failed for port 53ec4989-7198-480a-ab0a-351544564777, please check neutron logs for more information. [ 578.536853] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] [ 578.536853] env[62460]: INFO nova.compute.manager [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] Terminating instance [ 578.543192] env[62460]: DEBUG oslo_concurrency.lockutils [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Acquiring lock "refresh_cache-8a2699f3-3cc4-4ec3-899a-216e09ad086a" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 578.650425] env[62460]: DEBUG nova.network.neutron [req-e2ba2685-590d-42f9-a848-894da55787c9 req-7a712bcd-63ec-4499-a5b8-e3c6a07aa609 service nova] [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 578.668299] env[62460]: DEBUG oslo_concurrency.lockutils [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Acquiring lock "8c64f76d-cb18-41c5-9afa-e88af038f2b0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.668517] env[62460]: DEBUG oslo_concurrency.lockutils [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Lock "8c64f76d-cb18-41c5-9afa-e88af038f2b0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 578.689640] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7735deb7-0ce2-4775-ac1a-33f037ece6d5 tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Lock "ce5ed11a-0e24-47d7-bfc4-3cf4181bdae3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 41.571s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 578.813466] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f0d2215-fa92-4dbc-8dc6-05528b15a23e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.823254] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1ea20e0-a06d-45aa-a52f-3a023152b8f7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.855817] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4f17057-dacf-4297-a013-366e8218444e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.863736] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41a0ed52-381b-4363-aeee-94c9fe1ba92a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.878040] env[62460]: DEBUG nova.compute.provider_tree [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 578.912559] env[62460]: DEBUG oslo_concurrency.lockutils [req-ba2e6a99-58ee-4714-8864-6b837a99a8bd req-508a1ce4-d973-4e01-9b47-5c207cd7f3f9 service nova] Releasing lock "refresh_cache-a3f7edbe-4bba-4fff-9e62-99b7a85f971b" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 578.912960] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] Acquired lock "refresh_cache-a3f7edbe-4bba-4fff-9e62-99b7a85f971b" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 578.913130] env[62460]: DEBUG nova.network.neutron [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 578.996868] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] Releasing lock "refresh_cache-d52580bc-7c11-47f0-a6c4-12941c48d75c" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 578.996868] env[62460]: DEBUG nova.compute.manager [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 578.996868] env[62460]: DEBUG nova.compute.manager [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 578.997268] env[62460]: DEBUG nova.network.neutron [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 579.057131] env[62460]: DEBUG nova.network.neutron [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 579.154637] env[62460]: DEBUG oslo_concurrency.lockutils [req-e2ba2685-590d-42f9-a848-894da55787c9 req-7a712bcd-63ec-4499-a5b8-e3c6a07aa609 service nova] Releasing lock "refresh_cache-8a2699f3-3cc4-4ec3-899a-216e09ad086a" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 579.155366] env[62460]: DEBUG oslo_concurrency.lockutils [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Acquired lock "refresh_cache-8a2699f3-3cc4-4ec3-899a-216e09ad086a" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 579.156931] env[62460]: DEBUG nova.network.neutron [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 579.193114] env[62460]: DEBUG nova.compute.manager [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] [instance: a2747204-95c5-4200-8742-9ec39b4368c8] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 579.383351] env[62460]: DEBUG nova.scheduler.client.report [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 579.475443] env[62460]: DEBUG nova.network.neutron [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 579.563318] env[62460]: DEBUG nova.network.neutron [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 579.706516] env[62460]: DEBUG nova.network.neutron [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 579.755823] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.804174] env[62460]: DEBUG nova.network.neutron [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 579.838868] env[62460]: DEBUG oslo_concurrency.lockutils [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Acquiring lock "a24776d9-9950-4c83-9641-9675cabd5fd3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.839277] env[62460]: DEBUG oslo_concurrency.lockutils [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Lock "a24776d9-9950-4c83-9641-9675cabd5fd3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.003s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.892933] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.522s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 579.894416] env[62460]: DEBUG nova.compute.manager [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 579.896420] env[62460]: DEBUG oslo_concurrency.lockutils [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.966s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.898550] env[62460]: INFO nova.compute.claims [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 579.925785] env[62460]: DEBUG nova.network.neutron [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 580.035439] env[62460]: DEBUG nova.compute.manager [req-4553e297-2755-436b-8c5b-f5ccc70162dd req-586788d6-f152-49f0-85e8-8b51f5e2b76e service nova] [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] Received event network-vif-deleted-0978e846-4023-4bcd-867b-55313312ab35 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 580.059465] env[62460]: DEBUG nova.compute.manager [req-9553d335-80df-465f-ab75-7e7fbf86147b req-ec6e91b2-c282-49d2-9644-70fd3a9c49a0 service nova] [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] Received event network-vif-deleted-53ec4989-7198-480a-ab0a-351544564777 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 580.065788] env[62460]: INFO nova.compute.manager [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] [instance: d52580bc-7c11-47f0-a6c4-12941c48d75c] Took 1.07 seconds to deallocate network for instance. [ 580.307554] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] Releasing lock "refresh_cache-a3f7edbe-4bba-4fff-9e62-99b7a85f971b" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 580.307554] env[62460]: DEBUG nova.compute.manager [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 580.307554] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 580.307893] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d236c12e-7641-4f08-9480-1e5cc5039bc5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.318299] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8a611eb-493a-4d59-83a4-b9632376f1ef {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.344539] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a3f7edbe-4bba-4fff-9e62-99b7a85f971b could not be found. [ 580.345193] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 580.345193] env[62460]: INFO nova.compute.manager [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] Took 0.04 seconds to destroy the instance on the hypervisor. [ 580.345193] env[62460]: DEBUG oslo.service.loopingcall [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 580.345403] env[62460]: DEBUG nova.compute.manager [-] [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 580.345696] env[62460]: DEBUG nova.network.neutron [-] [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 580.377317] env[62460]: DEBUG nova.network.neutron [-] [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 580.403978] env[62460]: DEBUG nova.compute.utils [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 580.411148] env[62460]: DEBUG nova.compute.manager [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 580.412046] env[62460]: DEBUG nova.network.neutron [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 580.429892] env[62460]: DEBUG oslo_concurrency.lockutils [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Releasing lock "refresh_cache-8a2699f3-3cc4-4ec3-899a-216e09ad086a" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 580.430351] env[62460]: DEBUG nova.compute.manager [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 580.431840] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 580.432441] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ca609427-8ee2-42b4-8918-e0d56905fb62 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.448716] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80ea0fb7-92ac-4323-b528-ec025840ebd8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.471049] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8a2699f3-3cc4-4ec3-899a-216e09ad086a could not be found. [ 580.473879] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 580.473879] env[62460]: INFO nova.compute.manager [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] Took 0.04 seconds to destroy the instance on the hypervisor. [ 580.473879] env[62460]: DEBUG oslo.service.loopingcall [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 580.473879] env[62460]: DEBUG nova.compute.manager [-] [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 580.473879] env[62460]: DEBUG nova.network.neutron [-] [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 580.509770] env[62460]: DEBUG nova.network.neutron [-] [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 580.581162] env[62460]: DEBUG nova.policy [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5f69b3e3f50248ff8e57054294a04686', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0da056d93bdf40c39d6e82e457727ff6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 580.879249] env[62460]: DEBUG nova.network.neutron [-] [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 580.914433] env[62460]: DEBUG nova.compute.manager [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 581.013012] env[62460]: DEBUG nova.network.neutron [-] [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 581.119054] env[62460]: INFO nova.scheduler.client.report [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] Deleted allocations for instance d52580bc-7c11-47f0-a6c4-12941c48d75c [ 581.378204] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5372251d-c82a-458c-99c9-5b22f2d15447 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.385870] env[62460]: INFO nova.compute.manager [-] [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] Took 1.04 seconds to deallocate network for instance. [ 581.388592] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3cb732e-a51c-44fd-a208-d4f40afff2d3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.393227] env[62460]: DEBUG nova.compute.claims [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 581.394111] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 581.428766] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b40a90e9-5e7f-42c4-a717-19b46332307e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.435711] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c250b6c5-6588-4ba9-8b0c-ad4240ad124b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.450297] env[62460]: DEBUG nova.compute.provider_tree [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 581.519968] env[62460]: INFO nova.compute.manager [-] [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] Took 1.05 seconds to deallocate network for instance. [ 581.525742] env[62460]: DEBUG nova.compute.claims [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 581.526022] env[62460]: DEBUG oslo_concurrency.lockutils [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 581.631812] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a609887a-5e97-4d5f-9bd9-0f2bb8b7a8e8 tempest-ServerAddressesTestJSON-876383904 tempest-ServerAddressesTestJSON-876383904-project-member] Lock "d52580bc-7c11-47f0-a6c4-12941c48d75c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.139s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 581.883357] env[62460]: DEBUG nova.network.neutron [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] Successfully created port: e61d2785-3f7f-42cd-81ba-b8dde9aa7863 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 581.933948] env[62460]: DEBUG nova.compute.manager [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 581.955490] env[62460]: DEBUG nova.scheduler.client.report [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 581.972904] env[62460]: DEBUG nova.virt.hardware [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 581.973177] env[62460]: DEBUG nova.virt.hardware [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 581.973393] env[62460]: DEBUG nova.virt.hardware [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 581.973587] env[62460]: DEBUG nova.virt.hardware [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 581.973734] env[62460]: DEBUG nova.virt.hardware [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 581.973874] env[62460]: DEBUG nova.virt.hardware [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 581.976346] env[62460]: DEBUG nova.virt.hardware [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 581.976346] env[62460]: DEBUG nova.virt.hardware [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 581.976761] env[62460]: DEBUG nova.virt.hardware [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 581.977663] env[62460]: DEBUG nova.virt.hardware [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 581.978270] env[62460]: DEBUG nova.virt.hardware [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 581.978883] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b8cc6dc-3e92-4e0d-910c-4339fc3163f2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.995923] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36914bfe-265c-40a9-b607-9070aed69ad2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.137098] env[62460]: DEBUG nova.compute.manager [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 582.465062] env[62460]: DEBUG oslo_concurrency.lockutils [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.568s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 582.465062] env[62460]: DEBUG nova.compute.manager [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 582.467347] env[62460]: DEBUG oslo_concurrency.lockutils [None req-18fe540e-36dc-4e28-a3bd-6e40afcbc385 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.057s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 582.467855] env[62460]: DEBUG nova.objects.instance [None req-18fe540e-36dc-4e28-a3bd-6e40afcbc385 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Lazy-loading 'resources' on Instance uuid 44a2ac83-3fff-4958-a1d3-b3884b42c211 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 582.666034] env[62460]: DEBUG oslo_concurrency.lockutils [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 582.854188] env[62460]: DEBUG nova.compute.manager [None req-2f8a42f3-9d9d-4da5-8d76-7ba44dc288f7 tempest-ServerDiagnosticsV248Test-2008062123 tempest-ServerDiagnosticsV248Test-2008062123-project-admin] [instance: 00544d2a-1a15-4347-abe4-3641b5bf0cfe] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 582.855808] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-731e7a7a-7b35-4655-a944-1e25a2af621c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.864889] env[62460]: INFO nova.compute.manager [None req-2f8a42f3-9d9d-4da5-8d76-7ba44dc288f7 tempest-ServerDiagnosticsV248Test-2008062123 tempest-ServerDiagnosticsV248Test-2008062123-project-admin] [instance: 00544d2a-1a15-4347-abe4-3641b5bf0cfe] Retrieving diagnostics [ 582.866623] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b10bfc9d-d81a-408c-93b1-54f142bdc13a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.974963] env[62460]: DEBUG nova.compute.utils [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 582.981135] env[62460]: DEBUG nova.compute.manager [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 582.981412] env[62460]: DEBUG nova.network.neutron [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 583.125474] env[62460]: DEBUG nova.policy [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b8f440f68fba4dc89e91c301b0fe07b7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '619748a24fb4457aaea6b0c596e2932d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 583.275300] env[62460]: DEBUG oslo_concurrency.lockutils [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] Acquiring lock "61426715-7a38-475d-895a-0eb6d6040c66" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.275595] env[62460]: DEBUG oslo_concurrency.lockutils [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] Lock "61426715-7a38-475d-895a-0eb6d6040c66" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 583.434179] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a91071b0-e2ca-4d49-a6de-a03206ad1b8d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.446524] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1c7751f-d7b5-4d12-af5c-e4448e1e3db9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.487062] env[62460]: DEBUG nova.compute.manager [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 583.494393] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7da61cd8-04ba-4a77-9288-e7e73ec6697c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.502459] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36732f5f-b316-433d-812e-4e4ad2a0a656 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.520246] env[62460]: DEBUG nova.compute.provider_tree [None req-18fe540e-36dc-4e28-a3bd-6e40afcbc385 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 584.023316] env[62460]: DEBUG nova.scheduler.client.report [None req-18fe540e-36dc-4e28-a3bd-6e40afcbc385 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 584.258212] env[62460]: DEBUG nova.network.neutron [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] Successfully created port: a9063309-85d1-4e90-8756-635659c53ff6 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 584.500449] env[62460]: DEBUG nova.compute.manager [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 584.529414] env[62460]: DEBUG oslo_concurrency.lockutils [None req-18fe540e-36dc-4e28-a3bd-6e40afcbc385 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.062s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 584.535953] env[62460]: DEBUG nova.virt.hardware [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 584.535953] env[62460]: DEBUG nova.virt.hardware [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 584.535953] env[62460]: DEBUG nova.virt.hardware [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 584.536215] env[62460]: DEBUG nova.virt.hardware [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 584.536215] env[62460]: DEBUG nova.virt.hardware [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 584.536215] env[62460]: DEBUG nova.virt.hardware [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 584.536635] env[62460]: DEBUG nova.virt.hardware [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 584.536917] env[62460]: DEBUG nova.virt.hardware [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 584.537220] env[62460]: DEBUG nova.virt.hardware [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 584.537496] env[62460]: DEBUG nova.virt.hardware [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 584.537779] env[62460]: DEBUG nova.virt.hardware [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 584.539471] env[62460]: DEBUG oslo_concurrency.lockutils [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.844s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 584.541687] env[62460]: INFO nova.compute.claims [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 584.545639] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9afd70f2-10b2-42da-8ac2-47203cb407d5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.555279] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af2a3327-2042-49e3-b76c-2723bbe05e9b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.576668] env[62460]: INFO nova.scheduler.client.report [None req-18fe540e-36dc-4e28-a3bd-6e40afcbc385 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Deleted allocations for instance 44a2ac83-3fff-4958-a1d3-b3884b42c211 [ 584.858985] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5a702432-05a2-486c-ae26-233347bf963c tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Acquiring lock "00544d2a-1a15-4347-abe4-3641b5bf0cfe" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 584.859283] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5a702432-05a2-486c-ae26-233347bf963c tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Lock "00544d2a-1a15-4347-abe4-3641b5bf0cfe" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 584.859496] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5a702432-05a2-486c-ae26-233347bf963c tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Acquiring lock "00544d2a-1a15-4347-abe4-3641b5bf0cfe-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 584.859744] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5a702432-05a2-486c-ae26-233347bf963c tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Lock "00544d2a-1a15-4347-abe4-3641b5bf0cfe-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 584.859841] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5a702432-05a2-486c-ae26-233347bf963c tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Lock "00544d2a-1a15-4347-abe4-3641b5bf0cfe-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 584.861919] env[62460]: INFO nova.compute.manager [None req-5a702432-05a2-486c-ae26-233347bf963c tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] [instance: 00544d2a-1a15-4347-abe4-3641b5bf0cfe] Terminating instance [ 584.863828] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5a702432-05a2-486c-ae26-233347bf963c tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Acquiring lock "refresh_cache-00544d2a-1a15-4347-abe4-3641b5bf0cfe" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 584.864011] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5a702432-05a2-486c-ae26-233347bf963c tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Acquired lock "refresh_cache-00544d2a-1a15-4347-abe4-3641b5bf0cfe" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 584.864159] env[62460]: DEBUG nova.network.neutron [None req-5a702432-05a2-486c-ae26-233347bf963c tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] [instance: 00544d2a-1a15-4347-abe4-3641b5bf0cfe] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 585.092348] env[62460]: DEBUG oslo_concurrency.lockutils [None req-18fe540e-36dc-4e28-a3bd-6e40afcbc385 tempest-ServersAdmin275Test-988659854 tempest-ServersAdmin275Test-988659854-project-member] Lock "44a2ac83-3fff-4958-a1d3-b3884b42c211" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.434s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 585.416243] env[62460]: DEBUG nova.network.neutron [None req-5a702432-05a2-486c-ae26-233347bf963c tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] [instance: 00544d2a-1a15-4347-abe4-3641b5bf0cfe] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 585.611889] env[62460]: DEBUG nova.network.neutron [None req-5a702432-05a2-486c-ae26-233347bf963c tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] [instance: 00544d2a-1a15-4347-abe4-3641b5bf0cfe] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 585.913393] env[62460]: ERROR nova.compute.manager [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e61d2785-3f7f-42cd-81ba-b8dde9aa7863, please check neutron logs for more information. [ 585.913393] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 585.913393] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 585.913393] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 585.913393] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 585.913393] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 585.913393] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 585.913393] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 585.913393] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 585.913393] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 585.913393] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 585.913393] env[62460]: ERROR nova.compute.manager raise self.value [ 585.913393] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 585.913393] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 585.913393] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 585.913393] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 585.913962] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 585.913962] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 585.913962] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e61d2785-3f7f-42cd-81ba-b8dde9aa7863, please check neutron logs for more information. [ 585.913962] env[62460]: ERROR nova.compute.manager [ 585.913962] env[62460]: Traceback (most recent call last): [ 585.913962] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 585.913962] env[62460]: listener.cb(fileno) [ 585.913962] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 585.913962] env[62460]: result = function(*args, **kwargs) [ 585.913962] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 585.913962] env[62460]: return func(*args, **kwargs) [ 585.913962] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 585.913962] env[62460]: raise e [ 585.913962] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 585.913962] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 585.913962] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 585.913962] env[62460]: created_port_ids = self._update_ports_for_instance( [ 585.913962] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 585.913962] env[62460]: with excutils.save_and_reraise_exception(): [ 585.913962] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 585.913962] env[62460]: self.force_reraise() [ 585.913962] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 585.913962] env[62460]: raise self.value [ 585.913962] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 585.913962] env[62460]: updated_port = self._update_port( [ 585.913962] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 585.913962] env[62460]: _ensure_no_port_binding_failure(port) [ 585.913962] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 585.913962] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 585.914801] env[62460]: nova.exception.PortBindingFailed: Binding failed for port e61d2785-3f7f-42cd-81ba-b8dde9aa7863, please check neutron logs for more information. [ 585.914801] env[62460]: Removing descriptor: 18 [ 585.915222] env[62460]: ERROR nova.compute.manager [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e61d2785-3f7f-42cd-81ba-b8dde9aa7863, please check neutron logs for more information. [ 585.915222] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] Traceback (most recent call last): [ 585.915222] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 585.915222] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] yield resources [ 585.915222] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 585.915222] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] self.driver.spawn(context, instance, image_meta, [ 585.915222] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 585.915222] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] self._vmops.spawn(context, instance, image_meta, injected_files, [ 585.915222] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 585.915222] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] vm_ref = self.build_virtual_machine(instance, [ 585.915222] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 585.916140] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] vif_infos = vmwarevif.get_vif_info(self._session, [ 585.916140] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 585.916140] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] for vif in network_info: [ 585.916140] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 585.916140] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] return self._sync_wrapper(fn, *args, **kwargs) [ 585.916140] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 585.916140] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] self.wait() [ 585.916140] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 585.916140] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] self[:] = self._gt.wait() [ 585.916140] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 585.916140] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] return self._exit_event.wait() [ 585.916140] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 585.916140] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] result = hub.switch() [ 585.916836] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 585.916836] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] return self.greenlet.switch() [ 585.916836] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 585.916836] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] result = function(*args, **kwargs) [ 585.916836] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 585.916836] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] return func(*args, **kwargs) [ 585.916836] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 585.916836] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] raise e [ 585.916836] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 585.916836] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] nwinfo = self.network_api.allocate_for_instance( [ 585.916836] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 585.916836] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] created_port_ids = self._update_ports_for_instance( [ 585.916836] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 585.917916] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] with excutils.save_and_reraise_exception(): [ 585.917916] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 585.917916] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] self.force_reraise() [ 585.917916] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 585.917916] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] raise self.value [ 585.917916] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 585.917916] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] updated_port = self._update_port( [ 585.917916] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 585.917916] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] _ensure_no_port_binding_failure(port) [ 585.917916] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 585.917916] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] raise exception.PortBindingFailed(port_id=port['id']) [ 585.917916] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] nova.exception.PortBindingFailed: Binding failed for port e61d2785-3f7f-42cd-81ba-b8dde9aa7863, please check neutron logs for more information. [ 585.917916] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] [ 585.918893] env[62460]: INFO nova.compute.manager [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] Terminating instance [ 585.918893] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "refresh_cache-b51ddc3a-de97-4d51-adc9-e92b4fb0a150" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 585.918893] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquired lock "refresh_cache-b51ddc3a-de97-4d51-adc9-e92b4fb0a150" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 585.918893] env[62460]: DEBUG nova.network.neutron [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 585.926039] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15691c56-992b-4e11-b737-05a275f83e13 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.938262] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2675da2a-9f31-41f4-9bb6-39a6baa2f8ff {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.976032] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44f29dee-a304-442a-b58a-e0fcca69a064 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.984539] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b83d5bc4-9783-41a8-ba16-36a2d4336e62 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.999975] env[62460]: DEBUG nova.compute.provider_tree [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 586.115851] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5a702432-05a2-486c-ae26-233347bf963c tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Releasing lock "refresh_cache-00544d2a-1a15-4347-abe4-3641b5bf0cfe" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 586.116250] env[62460]: DEBUG nova.compute.manager [None req-5a702432-05a2-486c-ae26-233347bf963c tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] [instance: 00544d2a-1a15-4347-abe4-3641b5bf0cfe] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 586.117310] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-5a702432-05a2-486c-ae26-233347bf963c tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] [instance: 00544d2a-1a15-4347-abe4-3641b5bf0cfe] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 586.117310] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-203fb7f2-de80-4ac8-9987-cd1e3430eeb4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.127966] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a702432-05a2-486c-ae26-233347bf963c tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] [instance: 00544d2a-1a15-4347-abe4-3641b5bf0cfe] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 586.128294] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4e303bdd-663f-4b22-af2c-ff7a152d413d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.135922] env[62460]: DEBUG oslo_vmware.api [None req-5a702432-05a2-486c-ae26-233347bf963c tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Waiting for the task: (returnval){ [ 586.135922] env[62460]: value = "task-1313434" [ 586.135922] env[62460]: _type = "Task" [ 586.135922] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 586.144903] env[62460]: DEBUG oslo_vmware.api [None req-5a702432-05a2-486c-ae26-233347bf963c tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Task: {'id': task-1313434, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 586.197161] env[62460]: DEBUG oslo_concurrency.lockutils [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "b6efad50-aa2e-49f9-9ce7-5fead31db7a8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 586.197458] env[62460]: DEBUG oslo_concurrency.lockutils [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "b6efad50-aa2e-49f9-9ce7-5fead31db7a8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 586.453962] env[62460]: DEBUG nova.network.neutron [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 586.505038] env[62460]: DEBUG nova.scheduler.client.report [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 586.540544] env[62460]: DEBUG nova.compute.manager [req-5bf3de47-9886-40d2-84a9-56cc77bcd0f2 req-d1421e2c-cf2c-4bab-9684-074d1f6ca128 service nova] [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] Received event network-changed-e61d2785-3f7f-42cd-81ba-b8dde9aa7863 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 586.540734] env[62460]: DEBUG nova.compute.manager [req-5bf3de47-9886-40d2-84a9-56cc77bcd0f2 req-d1421e2c-cf2c-4bab-9684-074d1f6ca128 service nova] [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] Refreshing instance network info cache due to event network-changed-e61d2785-3f7f-42cd-81ba-b8dde9aa7863. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 586.540922] env[62460]: DEBUG oslo_concurrency.lockutils [req-5bf3de47-9886-40d2-84a9-56cc77bcd0f2 req-d1421e2c-cf2c-4bab-9684-074d1f6ca128 service nova] Acquiring lock "refresh_cache-b51ddc3a-de97-4d51-adc9-e92b4fb0a150" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 586.648794] env[62460]: DEBUG oslo_vmware.api [None req-5a702432-05a2-486c-ae26-233347bf963c tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Task: {'id': task-1313434, 'name': PowerOffVM_Task, 'duration_secs': 0.1289} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 586.649168] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a702432-05a2-486c-ae26-233347bf963c tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] [instance: 00544d2a-1a15-4347-abe4-3641b5bf0cfe] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 586.649326] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-5a702432-05a2-486c-ae26-233347bf963c tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] [instance: 00544d2a-1a15-4347-abe4-3641b5bf0cfe] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 586.649614] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-39d11c73-0a04-4841-acb6-c66f619dac80 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.675887] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-5a702432-05a2-486c-ae26-233347bf963c tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] [instance: 00544d2a-1a15-4347-abe4-3641b5bf0cfe] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 586.677132] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-5a702432-05a2-486c-ae26-233347bf963c tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] [instance: 00544d2a-1a15-4347-abe4-3641b5bf0cfe] Deleting contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 586.677132] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a702432-05a2-486c-ae26-233347bf963c tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Deleting the datastore file [datastore1] 00544d2a-1a15-4347-abe4-3641b5bf0cfe {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 586.677412] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-70d86727-cf6c-4971-b3ce-448a6eef9c4e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.684385] env[62460]: DEBUG oslo_vmware.api [None req-5a702432-05a2-486c-ae26-233347bf963c tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Waiting for the task: (returnval){ [ 586.684385] env[62460]: value = "task-1313436" [ 586.684385] env[62460]: _type = "Task" [ 586.684385] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 586.695220] env[62460]: DEBUG oslo_vmware.api [None req-5a702432-05a2-486c-ae26-233347bf963c tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Task: {'id': task-1313436, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 586.716977] env[62460]: DEBUG nova.network.neutron [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.009984] env[62460]: DEBUG oslo_concurrency.lockutils [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.470s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 587.012923] env[62460]: DEBUG nova.compute.manager [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 587.014534] env[62460]: DEBUG oslo_concurrency.lockutils [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.932s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 587.016180] env[62460]: INFO nova.compute.claims [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 587.164152] env[62460]: DEBUG oslo_concurrency.lockutils [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] Acquiring lock "252df6ad-e29d-4596-bc99-4aae144bbcc5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.167019] env[62460]: DEBUG oslo_concurrency.lockutils [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] Lock "252df6ad-e29d-4596-bc99-4aae144bbcc5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 587.198834] env[62460]: DEBUG oslo_vmware.api [None req-5a702432-05a2-486c-ae26-233347bf963c tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Task: {'id': task-1313436, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.151347} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 587.199121] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a702432-05a2-486c-ae26-233347bf963c tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 587.199319] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-5a702432-05a2-486c-ae26-233347bf963c tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] [instance: 00544d2a-1a15-4347-abe4-3641b5bf0cfe] Deleted contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 587.199479] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-5a702432-05a2-486c-ae26-233347bf963c tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] [instance: 00544d2a-1a15-4347-abe4-3641b5bf0cfe] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 587.199647] env[62460]: INFO nova.compute.manager [None req-5a702432-05a2-486c-ae26-233347bf963c tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] [instance: 00544d2a-1a15-4347-abe4-3641b5bf0cfe] Took 1.08 seconds to destroy the instance on the hypervisor. [ 587.199884] env[62460]: DEBUG oslo.service.loopingcall [None req-5a702432-05a2-486c-ae26-233347bf963c tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 587.200091] env[62460]: DEBUG nova.compute.manager [-] [instance: 00544d2a-1a15-4347-abe4-3641b5bf0cfe] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 587.200188] env[62460]: DEBUG nova.network.neutron [-] [instance: 00544d2a-1a15-4347-abe4-3641b5bf0cfe] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 587.223028] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Releasing lock "refresh_cache-b51ddc3a-de97-4d51-adc9-e92b4fb0a150" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 587.223028] env[62460]: DEBUG nova.compute.manager [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 587.223028] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 587.223028] env[62460]: DEBUG oslo_concurrency.lockutils [req-5bf3de47-9886-40d2-84a9-56cc77bcd0f2 req-d1421e2c-cf2c-4bab-9684-074d1f6ca128 service nova] Acquired lock "refresh_cache-b51ddc3a-de97-4d51-adc9-e92b4fb0a150" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 587.223028] env[62460]: DEBUG nova.network.neutron [req-5bf3de47-9886-40d2-84a9-56cc77bcd0f2 req-d1421e2c-cf2c-4bab-9684-074d1f6ca128 service nova] [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] Refreshing network info cache for port e61d2785-3f7f-42cd-81ba-b8dde9aa7863 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 587.223657] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a733236f-a8ca-4ea1-8834-32b00a27cfb2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.240272] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4468f114-f387-450f-8369-c06e8a08c8b6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.251166] env[62460]: DEBUG nova.network.neutron [-] [instance: 00544d2a-1a15-4347-abe4-3641b5bf0cfe] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 587.266190] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b51ddc3a-de97-4d51-adc9-e92b4fb0a150 could not be found. [ 587.266445] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 587.266625] env[62460]: INFO nova.compute.manager [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] Took 0.04 seconds to destroy the instance on the hypervisor. [ 587.266873] env[62460]: DEBUG oslo.service.loopingcall [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 587.267364] env[62460]: DEBUG nova.compute.manager [-] [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 587.267468] env[62460]: DEBUG nova.network.neutron [-] [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 587.299854] env[62460]: ERROR nova.compute.manager [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a9063309-85d1-4e90-8756-635659c53ff6, please check neutron logs for more information. [ 587.299854] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 587.299854] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 587.299854] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 587.299854] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 587.299854] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 587.299854] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 587.299854] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 587.299854] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 587.299854] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 587.299854] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 587.299854] env[62460]: ERROR nova.compute.manager raise self.value [ 587.299854] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 587.299854] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 587.299854] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 587.299854] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 587.300531] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 587.300531] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 587.300531] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a9063309-85d1-4e90-8756-635659c53ff6, please check neutron logs for more information. [ 587.300531] env[62460]: ERROR nova.compute.manager [ 587.300531] env[62460]: Traceback (most recent call last): [ 587.300531] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 587.300531] env[62460]: listener.cb(fileno) [ 587.300531] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 587.300531] env[62460]: result = function(*args, **kwargs) [ 587.300531] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 587.300531] env[62460]: return func(*args, **kwargs) [ 587.300531] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 587.300531] env[62460]: raise e [ 587.300531] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 587.300531] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 587.300531] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 587.300531] env[62460]: created_port_ids = self._update_ports_for_instance( [ 587.300531] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 587.300531] env[62460]: with excutils.save_and_reraise_exception(): [ 587.300531] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 587.300531] env[62460]: self.force_reraise() [ 587.300531] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 587.300531] env[62460]: raise self.value [ 587.300531] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 587.300531] env[62460]: updated_port = self._update_port( [ 587.300531] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 587.300531] env[62460]: _ensure_no_port_binding_failure(port) [ 587.300531] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 587.300531] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 587.301995] env[62460]: nova.exception.PortBindingFailed: Binding failed for port a9063309-85d1-4e90-8756-635659c53ff6, please check neutron logs for more information. [ 587.301995] env[62460]: Removing descriptor: 17 [ 587.301995] env[62460]: ERROR nova.compute.manager [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a9063309-85d1-4e90-8756-635659c53ff6, please check neutron logs for more information. [ 587.301995] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] Traceback (most recent call last): [ 587.301995] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 587.301995] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] yield resources [ 587.301995] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 587.301995] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] self.driver.spawn(context, instance, image_meta, [ 587.301995] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 587.301995] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] self._vmops.spawn(context, instance, image_meta, injected_files, [ 587.301995] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 587.301995] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] vm_ref = self.build_virtual_machine(instance, [ 587.302412] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 587.302412] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] vif_infos = vmwarevif.get_vif_info(self._session, [ 587.302412] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 587.302412] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] for vif in network_info: [ 587.302412] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 587.302412] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] return self._sync_wrapper(fn, *args, **kwargs) [ 587.302412] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 587.302412] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] self.wait() [ 587.302412] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 587.302412] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] self[:] = self._gt.wait() [ 587.302412] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 587.302412] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] return self._exit_event.wait() [ 587.302412] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 587.302790] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] result = hub.switch() [ 587.302790] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 587.302790] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] return self.greenlet.switch() [ 587.302790] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 587.302790] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] result = function(*args, **kwargs) [ 587.302790] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 587.302790] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] return func(*args, **kwargs) [ 587.302790] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 587.302790] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] raise e [ 587.302790] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 587.302790] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] nwinfo = self.network_api.allocate_for_instance( [ 587.302790] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 587.302790] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] created_port_ids = self._update_ports_for_instance( [ 587.305154] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 587.305154] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] with excutils.save_and_reraise_exception(): [ 587.305154] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 587.305154] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] self.force_reraise() [ 587.305154] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 587.305154] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] raise self.value [ 587.305154] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 587.305154] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] updated_port = self._update_port( [ 587.305154] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 587.305154] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] _ensure_no_port_binding_failure(port) [ 587.305154] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 587.305154] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] raise exception.PortBindingFailed(port_id=port['id']) [ 587.305698] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] nova.exception.PortBindingFailed: Binding failed for port a9063309-85d1-4e90-8756-635659c53ff6, please check neutron logs for more information. [ 587.305698] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] [ 587.305698] env[62460]: INFO nova.compute.manager [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] Terminating instance [ 587.305698] env[62460]: DEBUG oslo_concurrency.lockutils [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Acquiring lock "refresh_cache-6ea9a22a-d6ca-4e14-b83e-97074d4cb531" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 587.305698] env[62460]: DEBUG oslo_concurrency.lockutils [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Acquired lock "refresh_cache-6ea9a22a-d6ca-4e14-b83e-97074d4cb531" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 587.305698] env[62460]: DEBUG nova.network.neutron [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 587.306470] env[62460]: DEBUG nova.network.neutron [-] [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 587.518265] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Acquiring lock "02cb219f-989f-45b5-a9ba-d0b5a521cfa2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.518564] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Lock "02cb219f-989f-45b5-a9ba-d0b5a521cfa2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 587.520906] env[62460]: DEBUG nova.compute.utils [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 587.523444] env[62460]: DEBUG nova.compute.manager [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 587.523807] env[62460]: DEBUG nova.network.neutron [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 587.646822] env[62460]: DEBUG nova.policy [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bec75436785e4b78aed71acf2c090fe7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f3a8a64d95f44ba9b6229961eea891d8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 587.760126] env[62460]: DEBUG nova.network.neutron [-] [instance: 00544d2a-1a15-4347-abe4-3641b5bf0cfe] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.770993] env[62460]: DEBUG nova.network.neutron [req-5bf3de47-9886-40d2-84a9-56cc77bcd0f2 req-d1421e2c-cf2c-4bab-9684-074d1f6ca128 service nova] [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 587.811018] env[62460]: DEBUG nova.network.neutron [-] [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.842957] env[62460]: DEBUG nova.network.neutron [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 588.015714] env[62460]: DEBUG nova.network.neutron [req-5bf3de47-9886-40d2-84a9-56cc77bcd0f2 req-d1421e2c-cf2c-4bab-9684-074d1f6ca128 service nova] [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.025579] env[62460]: DEBUG nova.compute.manager [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 588.066746] env[62460]: DEBUG nova.network.neutron [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.262883] env[62460]: INFO nova.compute.manager [-] [instance: 00544d2a-1a15-4347-abe4-3641b5bf0cfe] Took 1.06 seconds to deallocate network for instance. [ 588.313229] env[62460]: INFO nova.compute.manager [-] [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] Took 1.05 seconds to deallocate network for instance. [ 588.319124] env[62460]: DEBUG nova.compute.claims [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 588.320080] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 588.371272] env[62460]: DEBUG nova.compute.manager [req-ec29b377-669a-4127-bf3d-8c0d595f9fbe req-5cd1b6e1-ec76-48f5-98d3-8797b971bcc1 service nova] [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] Received event network-changed-a9063309-85d1-4e90-8756-635659c53ff6 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 588.371272] env[62460]: DEBUG nova.compute.manager [req-ec29b377-669a-4127-bf3d-8c0d595f9fbe req-5cd1b6e1-ec76-48f5-98d3-8797b971bcc1 service nova] [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] Refreshing instance network info cache due to event network-changed-a9063309-85d1-4e90-8756-635659c53ff6. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 588.371272] env[62460]: DEBUG oslo_concurrency.lockutils [req-ec29b377-669a-4127-bf3d-8c0d595f9fbe req-5cd1b6e1-ec76-48f5-98d3-8797b971bcc1 service nova] Acquiring lock "refresh_cache-6ea9a22a-d6ca-4e14-b83e-97074d4cb531" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 588.412264] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] Acquiring lock "578627be-e695-4953-8d0e-9763d12b9a28" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 588.412491] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] Lock "578627be-e695-4953-8d0e-9763d12b9a28" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 588.521899] env[62460]: DEBUG oslo_concurrency.lockutils [req-5bf3de47-9886-40d2-84a9-56cc77bcd0f2 req-d1421e2c-cf2c-4bab-9684-074d1f6ca128 service nova] Releasing lock "refresh_cache-b51ddc3a-de97-4d51-adc9-e92b4fb0a150" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 588.542758] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0400373-6e20-4025-927a-f9ca85852ddb {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.550978] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b59485e1-2246-4d5c-8463-b8ad4f1a03f7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.593619] env[62460]: DEBUG oslo_concurrency.lockutils [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Releasing lock "refresh_cache-6ea9a22a-d6ca-4e14-b83e-97074d4cb531" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 588.593675] env[62460]: DEBUG nova.compute.manager [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 588.593863] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 588.597134] env[62460]: DEBUG nova.network.neutron [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] Successfully created port: fbb7d4a3-775b-4aac-aa13-f88699a16b64 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 588.599026] env[62460]: DEBUG oslo_concurrency.lockutils [req-ec29b377-669a-4127-bf3d-8c0d595f9fbe req-5cd1b6e1-ec76-48f5-98d3-8797b971bcc1 service nova] Acquired lock "refresh_cache-6ea9a22a-d6ca-4e14-b83e-97074d4cb531" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 588.599026] env[62460]: DEBUG nova.network.neutron [req-ec29b377-669a-4127-bf3d-8c0d595f9fbe req-5cd1b6e1-ec76-48f5-98d3-8797b971bcc1 service nova] [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] Refreshing network info cache for port a9063309-85d1-4e90-8756-635659c53ff6 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 588.599978] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b9f191aa-d05d-4f9b-9afc-ce0407212982 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.602611] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa35d000-59e7-4b38-8b06-62679d38bf27 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.612173] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82999582-7a53-4255-a600-b7bffaa378d3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.619115] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bcb2e77-ef53-4bc8-b222-fadebda18b5e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.640506] env[62460]: DEBUG nova.compute.provider_tree [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 588.649000] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6ea9a22a-d6ca-4e14-b83e-97074d4cb531 could not be found. [ 588.649000] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 588.649000] env[62460]: INFO nova.compute.manager [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] Took 0.05 seconds to destroy the instance on the hypervisor. [ 588.649000] env[62460]: DEBUG oslo.service.loopingcall [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 588.649000] env[62460]: DEBUG nova.compute.manager [-] [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 588.649000] env[62460]: DEBUG nova.network.neutron [-] [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 588.670877] env[62460]: DEBUG nova.network.neutron [-] [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 588.776333] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5a702432-05a2-486c-ae26-233347bf963c tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 588.920818] env[62460]: DEBUG oslo_concurrency.lockutils [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] Acquiring lock "724a8083-2bde-483e-bd5a-a928def284bf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 588.921115] env[62460]: DEBUG oslo_concurrency.lockutils [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] Lock "724a8083-2bde-483e-bd5a-a928def284bf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 589.019618] env[62460]: DEBUG nova.compute.manager [req-786712d2-1df8-4290-9bc1-c68b90c571a8 req-096d6717-01da-48f5-beee-3b25d082015e service nova] [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] Received event network-vif-deleted-e61d2785-3f7f-42cd-81ba-b8dde9aa7863 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 589.043818] env[62460]: DEBUG nova.compute.manager [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 589.080017] env[62460]: DEBUG nova.virt.hardware [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 589.080017] env[62460]: DEBUG nova.virt.hardware [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 589.080017] env[62460]: DEBUG nova.virt.hardware [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 589.080310] env[62460]: DEBUG nova.virt.hardware [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 589.080310] env[62460]: DEBUG nova.virt.hardware [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 589.080310] env[62460]: DEBUG nova.virt.hardware [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 589.080310] env[62460]: DEBUG nova.virt.hardware [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 589.080310] env[62460]: DEBUG nova.virt.hardware [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 589.080552] env[62460]: DEBUG nova.virt.hardware [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 589.080552] env[62460]: DEBUG nova.virt.hardware [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 589.080552] env[62460]: DEBUG nova.virt.hardware [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 589.080552] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fc74c56-6ec2-44b1-b767-a75cc0763559 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.090242] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c4b0b49-b971-4c25-82d5-eb5ed34d32ae {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.137107] env[62460]: DEBUG nova.network.neutron [req-ec29b377-669a-4127-bf3d-8c0d595f9fbe req-5cd1b6e1-ec76-48f5-98d3-8797b971bcc1 service nova] [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 589.144556] env[62460]: DEBUG nova.scheduler.client.report [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 589.173713] env[62460]: DEBUG nova.network.neutron [-] [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 589.257844] env[62460]: DEBUG nova.network.neutron [req-ec29b377-669a-4127-bf3d-8c0d595f9fbe req-5cd1b6e1-ec76-48f5-98d3-8797b971bcc1 service nova] [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 589.651772] env[62460]: DEBUG oslo_concurrency.lockutils [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.637s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 589.654217] env[62460]: DEBUG nova.compute.manager [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 589.656330] env[62460]: DEBUG oslo_concurrency.lockutils [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.428s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 589.658503] env[62460]: INFO nova.compute.claims [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 589.676674] env[62460]: INFO nova.compute.manager [-] [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] Took 1.03 seconds to deallocate network for instance. [ 589.680777] env[62460]: DEBUG nova.compute.claims [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 589.680777] env[62460]: DEBUG oslo_concurrency.lockutils [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.763352] env[62460]: DEBUG oslo_concurrency.lockutils [req-ec29b377-669a-4127-bf3d-8c0d595f9fbe req-5cd1b6e1-ec76-48f5-98d3-8797b971bcc1 service nova] Releasing lock "refresh_cache-6ea9a22a-d6ca-4e14-b83e-97074d4cb531" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 590.086409] env[62460]: ERROR nova.compute.manager [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port fbb7d4a3-775b-4aac-aa13-f88699a16b64, please check neutron logs for more information. [ 590.086409] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 590.086409] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 590.086409] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 590.086409] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 590.086409] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 590.086409] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 590.086409] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 590.086409] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 590.086409] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 590.086409] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 590.086409] env[62460]: ERROR nova.compute.manager raise self.value [ 590.086409] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 590.086409] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 590.086409] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 590.086409] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 590.087033] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 590.087033] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 590.087033] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port fbb7d4a3-775b-4aac-aa13-f88699a16b64, please check neutron logs for more information. [ 590.087033] env[62460]: ERROR nova.compute.manager [ 590.087033] env[62460]: Traceback (most recent call last): [ 590.087033] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 590.087033] env[62460]: listener.cb(fileno) [ 590.087033] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 590.087033] env[62460]: result = function(*args, **kwargs) [ 590.087033] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 590.087033] env[62460]: return func(*args, **kwargs) [ 590.087033] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 590.087033] env[62460]: raise e [ 590.087033] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 590.087033] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 590.087033] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 590.087033] env[62460]: created_port_ids = self._update_ports_for_instance( [ 590.087033] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 590.087033] env[62460]: with excutils.save_and_reraise_exception(): [ 590.087033] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 590.087033] env[62460]: self.force_reraise() [ 590.087033] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 590.087033] env[62460]: raise self.value [ 590.087033] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 590.087033] env[62460]: updated_port = self._update_port( [ 590.087033] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 590.087033] env[62460]: _ensure_no_port_binding_failure(port) [ 590.087033] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 590.087033] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 590.087742] env[62460]: nova.exception.PortBindingFailed: Binding failed for port fbb7d4a3-775b-4aac-aa13-f88699a16b64, please check neutron logs for more information. [ 590.087742] env[62460]: Removing descriptor: 18 [ 590.087742] env[62460]: ERROR nova.compute.manager [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port fbb7d4a3-775b-4aac-aa13-f88699a16b64, please check neutron logs for more information. [ 590.087742] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] Traceback (most recent call last): [ 590.087742] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 590.087742] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] yield resources [ 590.087742] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 590.087742] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] self.driver.spawn(context, instance, image_meta, [ 590.087742] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 590.087742] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] self._vmops.spawn(context, instance, image_meta, injected_files, [ 590.087742] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 590.087742] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] vm_ref = self.build_virtual_machine(instance, [ 590.088040] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 590.088040] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] vif_infos = vmwarevif.get_vif_info(self._session, [ 590.088040] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 590.088040] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] for vif in network_info: [ 590.088040] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 590.088040] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] return self._sync_wrapper(fn, *args, **kwargs) [ 590.088040] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 590.088040] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] self.wait() [ 590.088040] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 590.088040] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] self[:] = self._gt.wait() [ 590.088040] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 590.088040] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] return self._exit_event.wait() [ 590.088040] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 590.088349] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] result = hub.switch() [ 590.088349] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 590.088349] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] return self.greenlet.switch() [ 590.088349] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 590.088349] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] result = function(*args, **kwargs) [ 590.088349] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 590.088349] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] return func(*args, **kwargs) [ 590.088349] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 590.088349] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] raise e [ 590.088349] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 590.088349] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] nwinfo = self.network_api.allocate_for_instance( [ 590.088349] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 590.088349] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] created_port_ids = self._update_ports_for_instance( [ 590.088685] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 590.088685] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] with excutils.save_and_reraise_exception(): [ 590.088685] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 590.088685] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] self.force_reraise() [ 590.088685] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 590.088685] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] raise self.value [ 590.088685] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 590.088685] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] updated_port = self._update_port( [ 590.088685] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 590.088685] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] _ensure_no_port_binding_failure(port) [ 590.088685] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 590.088685] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] raise exception.PortBindingFailed(port_id=port['id']) [ 590.088966] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] nova.exception.PortBindingFailed: Binding failed for port fbb7d4a3-775b-4aac-aa13-f88699a16b64, please check neutron logs for more information. [ 590.088966] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] [ 590.088966] env[62460]: INFO nova.compute.manager [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] Terminating instance [ 590.090365] env[62460]: DEBUG oslo_concurrency.lockutils [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] Acquiring lock "refresh_cache-291beb16-db0e-4eb6-a224-2acfc14d6d77" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 590.090650] env[62460]: DEBUG oslo_concurrency.lockutils [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] Acquired lock "refresh_cache-291beb16-db0e-4eb6-a224-2acfc14d6d77" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 590.090650] env[62460]: DEBUG nova.network.neutron [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 590.165529] env[62460]: DEBUG nova.compute.utils [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 590.173179] env[62460]: DEBUG nova.compute.manager [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 590.173179] env[62460]: DEBUG nova.network.neutron [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 590.253887] env[62460]: DEBUG nova.policy [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b8f440f68fba4dc89e91c301b0fe07b7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '619748a24fb4457aaea6b0c596e2932d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 590.360685] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Acquiring lock "0d29a4ce-6689-4d0c-8144-e05853eb3a60" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 590.360824] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Lock "0d29a4ce-6689-4d0c-8144-e05853eb3a60" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 590.619461] env[62460]: DEBUG nova.network.neutron [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 590.644948] env[62460]: DEBUG nova.network.neutron [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] Successfully created port: 14e5de25-c06c-48c1-a690-db4605dc0193 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 590.673569] env[62460]: DEBUG nova.compute.manager [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 590.799263] env[62460]: DEBUG nova.network.neutron [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 590.969213] env[62460]: DEBUG nova.compute.manager [req-786bac3c-1a34-4c14-8ab6-9f66d99b6d31 req-7aef2fd5-142e-4641-baec-03341d0f25c3 service nova] [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] Received event network-vif-deleted-a9063309-85d1-4e90-8756-635659c53ff6 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 591.136379] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-858c1c44-1fb0-436e-924d-a972c5119ea9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.145341] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1547a7e9-2b6e-4160-bb60-01cd47ab74e1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.182167] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f4716fc-907f-42c5-a548-e86cdc4ecdfa {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.192522] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f2ddec5-624a-4223-9ce9-97c027dfcbd1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.206652] env[62460]: DEBUG nova.compute.provider_tree [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 591.305891] env[62460]: DEBUG oslo_concurrency.lockutils [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] Releasing lock "refresh_cache-291beb16-db0e-4eb6-a224-2acfc14d6d77" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 591.306501] env[62460]: DEBUG nova.compute.manager [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 591.307213] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 591.308449] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a82ea3f0-f85f-4028-8ce5-b3e2ab444f85 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.317628] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d656b53e-f015-46dd-ac78-2fc59735ec0b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.343013] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 291beb16-db0e-4eb6-a224-2acfc14d6d77 could not be found. [ 591.343013] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 591.343013] env[62460]: INFO nova.compute.manager [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] Took 0.03 seconds to destroy the instance on the hypervisor. [ 591.343013] env[62460]: DEBUG oslo.service.loopingcall [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 591.343013] env[62460]: DEBUG nova.compute.manager [-] [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 591.343013] env[62460]: DEBUG nova.network.neutron [-] [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 591.455102] env[62460]: DEBUG nova.compute.manager [req-3b5d1dd8-22ae-47a3-91e9-ded0f472289c req-862c04a1-cb4b-4d17-8ea5-bfaa83d02c15 service nova] [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] Received event network-changed-fbb7d4a3-775b-4aac-aa13-f88699a16b64 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 591.455302] env[62460]: DEBUG nova.compute.manager [req-3b5d1dd8-22ae-47a3-91e9-ded0f472289c req-862c04a1-cb4b-4d17-8ea5-bfaa83d02c15 service nova] [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] Refreshing instance network info cache due to event network-changed-fbb7d4a3-775b-4aac-aa13-f88699a16b64. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 591.455520] env[62460]: DEBUG oslo_concurrency.lockutils [req-3b5d1dd8-22ae-47a3-91e9-ded0f472289c req-862c04a1-cb4b-4d17-8ea5-bfaa83d02c15 service nova] Acquiring lock "refresh_cache-291beb16-db0e-4eb6-a224-2acfc14d6d77" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 591.455665] env[62460]: DEBUG oslo_concurrency.lockutils [req-3b5d1dd8-22ae-47a3-91e9-ded0f472289c req-862c04a1-cb4b-4d17-8ea5-bfaa83d02c15 service nova] Acquired lock "refresh_cache-291beb16-db0e-4eb6-a224-2acfc14d6d77" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 591.455823] env[62460]: DEBUG nova.network.neutron [req-3b5d1dd8-22ae-47a3-91e9-ded0f472289c req-862c04a1-cb4b-4d17-8ea5-bfaa83d02c15 service nova] [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] Refreshing network info cache for port fbb7d4a3-775b-4aac-aa13-f88699a16b64 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 591.522736] env[62460]: DEBUG nova.network.neutron [-] [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 591.610912] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Acquiring lock "02b7a3c3-bdfd-47d9-acf7-afd7725cea64" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.611218] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Lock "02b7a3c3-bdfd-47d9-acf7-afd7725cea64" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 591.688391] env[62460]: DEBUG nova.compute.manager [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 591.712520] env[62460]: DEBUG nova.scheduler.client.report [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 591.729508] env[62460]: DEBUG nova.virt.hardware [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 591.730055] env[62460]: DEBUG nova.virt.hardware [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 591.731137] env[62460]: DEBUG nova.virt.hardware [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 591.731137] env[62460]: DEBUG nova.virt.hardware [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 591.731137] env[62460]: DEBUG nova.virt.hardware [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 591.731137] env[62460]: DEBUG nova.virt.hardware [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 591.731137] env[62460]: DEBUG nova.virt.hardware [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 591.731456] env[62460]: DEBUG nova.virt.hardware [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 591.731687] env[62460]: DEBUG nova.virt.hardware [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 591.732046] env[62460]: DEBUG nova.virt.hardware [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 591.732280] env[62460]: DEBUG nova.virt.hardware [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 591.733237] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55fb0749-5309-4427-bb8d-d752872d28b2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.743145] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7df21655-1041-49e2-84d2-5f6b7b67b3cb {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.845224] env[62460]: ERROR nova.compute.manager [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 14e5de25-c06c-48c1-a690-db4605dc0193, please check neutron logs for more information. [ 591.845224] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 591.845224] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 591.845224] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 591.845224] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 591.845224] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 591.845224] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 591.845224] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 591.845224] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 591.845224] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 591.845224] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 591.845224] env[62460]: ERROR nova.compute.manager raise self.value [ 591.845224] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 591.845224] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 591.845224] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 591.845224] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 591.845571] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 591.845571] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 591.845571] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 14e5de25-c06c-48c1-a690-db4605dc0193, please check neutron logs for more information. [ 591.845571] env[62460]: ERROR nova.compute.manager [ 591.845571] env[62460]: Traceback (most recent call last): [ 591.845571] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 591.845571] env[62460]: listener.cb(fileno) [ 591.845571] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 591.845571] env[62460]: result = function(*args, **kwargs) [ 591.845571] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 591.845571] env[62460]: return func(*args, **kwargs) [ 591.845571] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 591.845571] env[62460]: raise e [ 591.845571] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 591.845571] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 591.845571] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 591.845571] env[62460]: created_port_ids = self._update_ports_for_instance( [ 591.845571] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 591.845571] env[62460]: with excutils.save_and_reraise_exception(): [ 591.845571] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 591.845571] env[62460]: self.force_reraise() [ 591.845571] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 591.845571] env[62460]: raise self.value [ 591.845571] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 591.845571] env[62460]: updated_port = self._update_port( [ 591.845571] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 591.845571] env[62460]: _ensure_no_port_binding_failure(port) [ 591.845571] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 591.845571] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 591.846140] env[62460]: nova.exception.PortBindingFailed: Binding failed for port 14e5de25-c06c-48c1-a690-db4605dc0193, please check neutron logs for more information. [ 591.846140] env[62460]: Removing descriptor: 18 [ 591.846140] env[62460]: ERROR nova.compute.manager [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 14e5de25-c06c-48c1-a690-db4605dc0193, please check neutron logs for more information. [ 591.846140] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] Traceback (most recent call last): [ 591.846140] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 591.846140] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] yield resources [ 591.846140] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 591.846140] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] self.driver.spawn(context, instance, image_meta, [ 591.846140] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 591.846140] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] self._vmops.spawn(context, instance, image_meta, injected_files, [ 591.846140] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 591.846140] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] vm_ref = self.build_virtual_machine(instance, [ 591.846384] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 591.846384] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] vif_infos = vmwarevif.get_vif_info(self._session, [ 591.846384] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 591.846384] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] for vif in network_info: [ 591.846384] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 591.846384] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] return self._sync_wrapper(fn, *args, **kwargs) [ 591.846384] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 591.846384] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] self.wait() [ 591.846384] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 591.846384] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] self[:] = self._gt.wait() [ 591.846384] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 591.846384] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] return self._exit_event.wait() [ 591.846384] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 591.846642] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] result = hub.switch() [ 591.846642] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 591.846642] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] return self.greenlet.switch() [ 591.846642] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 591.846642] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] result = function(*args, **kwargs) [ 591.846642] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 591.846642] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] return func(*args, **kwargs) [ 591.846642] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 591.846642] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] raise e [ 591.846642] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 591.846642] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] nwinfo = self.network_api.allocate_for_instance( [ 591.846642] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 591.846642] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] created_port_ids = self._update_ports_for_instance( [ 591.846900] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 591.846900] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] with excutils.save_and_reraise_exception(): [ 591.846900] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 591.846900] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] self.force_reraise() [ 591.846900] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 591.846900] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] raise self.value [ 591.846900] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 591.846900] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] updated_port = self._update_port( [ 591.846900] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 591.846900] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] _ensure_no_port_binding_failure(port) [ 591.846900] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 591.846900] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] raise exception.PortBindingFailed(port_id=port['id']) [ 591.847210] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] nova.exception.PortBindingFailed: Binding failed for port 14e5de25-c06c-48c1-a690-db4605dc0193, please check neutron logs for more information. [ 591.847210] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] [ 591.847210] env[62460]: INFO nova.compute.manager [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] Terminating instance [ 591.848798] env[62460]: DEBUG oslo_concurrency.lockutils [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Acquiring lock "refresh_cache-9b8ff88e-9e13-42b2-92b0-13af6873c916" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 591.849011] env[62460]: DEBUG oslo_concurrency.lockutils [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Acquired lock "refresh_cache-9b8ff88e-9e13-42b2-92b0-13af6873c916" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 591.849230] env[62460]: DEBUG nova.network.neutron [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 591.976688] env[62460]: DEBUG nova.network.neutron [req-3b5d1dd8-22ae-47a3-91e9-ded0f472289c req-862c04a1-cb4b-4d17-8ea5-bfaa83d02c15 service nova] [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 592.028509] env[62460]: DEBUG nova.network.neutron [-] [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 592.108361] env[62460]: DEBUG nova.network.neutron [req-3b5d1dd8-22ae-47a3-91e9-ded0f472289c req-862c04a1-cb4b-4d17-8ea5-bfaa83d02c15 service nova] [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 592.222622] env[62460]: DEBUG oslo_concurrency.lockutils [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.566s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 592.223092] env[62460]: DEBUG nova.compute.manager [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 592.225798] env[62460]: DEBUG oslo_concurrency.lockutils [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.590s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 592.372170] env[62460]: DEBUG nova.network.neutron [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 592.480858] env[62460]: DEBUG nova.network.neutron [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 592.533916] env[62460]: INFO nova.compute.manager [-] [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] Took 1.19 seconds to deallocate network for instance. [ 592.536609] env[62460]: DEBUG nova.compute.claims [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 592.536896] env[62460]: DEBUG oslo_concurrency.lockutils [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.610212] env[62460]: DEBUG oslo_concurrency.lockutils [req-3b5d1dd8-22ae-47a3-91e9-ded0f472289c req-862c04a1-cb4b-4d17-8ea5-bfaa83d02c15 service nova] Releasing lock "refresh_cache-291beb16-db0e-4eb6-a224-2acfc14d6d77" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 592.610475] env[62460]: DEBUG nova.compute.manager [req-3b5d1dd8-22ae-47a3-91e9-ded0f472289c req-862c04a1-cb4b-4d17-8ea5-bfaa83d02c15 service nova] [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] Received event network-vif-deleted-fbb7d4a3-775b-4aac-aa13-f88699a16b64 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 592.732862] env[62460]: DEBUG nova.compute.utils [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 592.738111] env[62460]: DEBUG nova.compute.manager [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 592.738111] env[62460]: DEBUG nova.network.neutron [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 592.972361] env[62460]: DEBUG nova.policy [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1eb34fe05d864fcfb9e0106f45ac82d7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ecfd8191d9f84e599ad9fbc90899d632', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 592.985828] env[62460]: DEBUG oslo_concurrency.lockutils [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Releasing lock "refresh_cache-9b8ff88e-9e13-42b2-92b0-13af6873c916" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 592.986287] env[62460]: DEBUG nova.compute.manager [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 592.986762] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 592.989580] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-87975594-e4c2-4dd4-8eb4-b7ff76e0aff5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.998245] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f4893f8-f0ee-4cd5-804e-75e810a16a69 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.026809] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9b8ff88e-9e13-42b2-92b0-13af6873c916 could not be found. [ 593.027110] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 593.027337] env[62460]: INFO nova.compute.manager [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] Took 0.04 seconds to destroy the instance on the hypervisor. [ 593.028321] env[62460]: DEBUG oslo.service.loopingcall [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 593.028321] env[62460]: DEBUG nova.compute.manager [-] [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 593.028321] env[62460]: DEBUG nova.network.neutron [-] [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 593.060141] env[62460]: DEBUG nova.network.neutron [-] [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 593.204174] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd4bf978-584b-4604-bc9f-5e6ffe2ceea7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.213021] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccfbb2e0-2c72-4dee-886e-6f1bc86fc405 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.245057] env[62460]: DEBUG nova.compute.manager [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 593.248262] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe63ce43-b5e1-4233-b56b-eec46c541d0c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.256511] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b48429b1-a5f7-40f2-a37f-bdb494cbd050 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.273018] env[62460]: DEBUG nova.compute.provider_tree [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 593.353916] env[62460]: DEBUG nova.network.neutron [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] Successfully created port: 536f1463-299b-4c7b-8962-741db114114c {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 593.564105] env[62460]: DEBUG nova.network.neutron [-] [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.567831] env[62460]: DEBUG nova.compute.manager [req-b61e841c-3a62-4198-9354-2a7b03df7f96 req-a783c17a-5c60-4bb2-91d2-939ab68141c5 service nova] [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] Received event network-changed-14e5de25-c06c-48c1-a690-db4605dc0193 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 593.568088] env[62460]: DEBUG nova.compute.manager [req-b61e841c-3a62-4198-9354-2a7b03df7f96 req-a783c17a-5c60-4bb2-91d2-939ab68141c5 service nova] [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] Refreshing instance network info cache due to event network-changed-14e5de25-c06c-48c1-a690-db4605dc0193. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 593.568380] env[62460]: DEBUG oslo_concurrency.lockutils [req-b61e841c-3a62-4198-9354-2a7b03df7f96 req-a783c17a-5c60-4bb2-91d2-939ab68141c5 service nova] Acquiring lock "refresh_cache-9b8ff88e-9e13-42b2-92b0-13af6873c916" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 593.568571] env[62460]: DEBUG oslo_concurrency.lockutils [req-b61e841c-3a62-4198-9354-2a7b03df7f96 req-a783c17a-5c60-4bb2-91d2-939ab68141c5 service nova] Acquired lock "refresh_cache-9b8ff88e-9e13-42b2-92b0-13af6873c916" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 593.568896] env[62460]: DEBUG nova.network.neutron [req-b61e841c-3a62-4198-9354-2a7b03df7f96 req-a783c17a-5c60-4bb2-91d2-939ab68141c5 service nova] [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] Refreshing network info cache for port 14e5de25-c06c-48c1-a690-db4605dc0193 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 593.775020] env[62460]: DEBUG nova.scheduler.client.report [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 594.072139] env[62460]: INFO nova.compute.manager [-] [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] Took 1.04 seconds to deallocate network for instance. [ 594.077899] env[62460]: DEBUG nova.compute.claims [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 594.078225] env[62460]: DEBUG oslo_concurrency.lockutils [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 594.107971] env[62460]: DEBUG nova.network.neutron [req-b61e841c-3a62-4198-9354-2a7b03df7f96 req-a783c17a-5c60-4bb2-91d2-939ab68141c5 service nova] [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 594.258625] env[62460]: DEBUG nova.compute.manager [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 594.279966] env[62460]: DEBUG oslo_concurrency.lockutils [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.054s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 594.280806] env[62460]: ERROR nova.compute.manager [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 357e567f-881f-453c-b117-bbf4d09e4a2f, please check neutron logs for more information. [ 594.280806] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] Traceback (most recent call last): [ 594.280806] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 594.280806] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] self.driver.spawn(context, instance, image_meta, [ 594.280806] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 594.280806] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] self._vmops.spawn(context, instance, image_meta, injected_files, [ 594.280806] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 594.280806] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] vm_ref = self.build_virtual_machine(instance, [ 594.280806] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 594.280806] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] vif_infos = vmwarevif.get_vif_info(self._session, [ 594.280806] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 594.281170] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] for vif in network_info: [ 594.281170] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 594.281170] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] return self._sync_wrapper(fn, *args, **kwargs) [ 594.281170] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 594.281170] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] self.wait() [ 594.281170] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 594.281170] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] self[:] = self._gt.wait() [ 594.281170] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 594.281170] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] return self._exit_event.wait() [ 594.281170] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 594.281170] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] result = hub.switch() [ 594.281170] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 594.281170] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] return self.greenlet.switch() [ 594.281468] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 594.281468] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] result = function(*args, **kwargs) [ 594.281468] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 594.281468] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] return func(*args, **kwargs) [ 594.281468] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 594.281468] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] raise e [ 594.281468] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 594.281468] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] nwinfo = self.network_api.allocate_for_instance( [ 594.281468] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 594.281468] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] created_port_ids = self._update_ports_for_instance( [ 594.281468] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 594.281468] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] with excutils.save_and_reraise_exception(): [ 594.281468] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 594.281792] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] self.force_reraise() [ 594.281792] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 594.281792] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] raise self.value [ 594.281792] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 594.281792] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] updated_port = self._update_port( [ 594.281792] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 594.281792] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] _ensure_no_port_binding_failure(port) [ 594.281792] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 594.281792] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] raise exception.PortBindingFailed(port_id=port['id']) [ 594.281792] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] nova.exception.PortBindingFailed: Binding failed for port 357e567f-881f-453c-b117-bbf4d09e4a2f, please check neutron logs for more information. [ 594.281792] env[62460]: ERROR nova.compute.manager [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] [ 594.284808] env[62460]: DEBUG nova.compute.utils [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] Binding failed for port 357e567f-881f-453c-b117-bbf4d09e4a2f, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 594.286024] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.531s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 594.290160] env[62460]: INFO nova.compute.claims [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] [instance: a2747204-95c5-4200-8742-9ec39b4368c8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 594.295863] env[62460]: DEBUG nova.virt.hardware [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 594.295863] env[62460]: DEBUG nova.virt.hardware [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 594.295863] env[62460]: DEBUG nova.virt.hardware [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 594.295863] env[62460]: DEBUG nova.virt.hardware [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 594.296931] env[62460]: DEBUG nova.virt.hardware [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 594.297338] env[62460]: DEBUG nova.virt.hardware [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 594.297621] env[62460]: DEBUG nova.virt.hardware [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 594.297973] env[62460]: DEBUG nova.virt.hardware [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 594.298257] env[62460]: DEBUG nova.virt.hardware [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 594.298467] env[62460]: DEBUG nova.virt.hardware [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 594.298682] env[62460]: DEBUG nova.virt.hardware [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 594.299228] env[62460]: DEBUG nova.compute.manager [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] Build of instance efef4327-22c8-4b59-9b26-89bcfe89bc59 was re-scheduled: Binding failed for port 357e567f-881f-453c-b117-bbf4d09e4a2f, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 594.300067] env[62460]: DEBUG nova.compute.manager [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 594.300067] env[62460]: DEBUG oslo_concurrency.lockutils [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Acquiring lock "refresh_cache-efef4327-22c8-4b59-9b26-89bcfe89bc59" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 594.300067] env[62460]: DEBUG oslo_concurrency.lockutils [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Acquired lock "refresh_cache-efef4327-22c8-4b59-9b26-89bcfe89bc59" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 594.300271] env[62460]: DEBUG nova.network.neutron [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 594.305109] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65445ad2-8edf-431f-9264-93896f0c4ee0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.307602] env[62460]: DEBUG nova.network.neutron [req-b61e841c-3a62-4198-9354-2a7b03df7f96 req-a783c17a-5c60-4bb2-91d2-939ab68141c5 service nova] [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.312765] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b35c48e-515e-4000-a6c1-1271d28919a4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.473610] env[62460]: ERROR nova.compute.manager [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 536f1463-299b-4c7b-8962-741db114114c, please check neutron logs for more information. [ 594.473610] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 594.473610] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 594.473610] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 594.473610] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 594.473610] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 594.473610] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 594.473610] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 594.473610] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 594.473610] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 594.473610] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 594.473610] env[62460]: ERROR nova.compute.manager raise self.value [ 594.473610] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 594.473610] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 594.473610] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 594.473610] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 594.476175] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 594.476175] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 594.476175] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 536f1463-299b-4c7b-8962-741db114114c, please check neutron logs for more information. [ 594.476175] env[62460]: ERROR nova.compute.manager [ 594.476175] env[62460]: Traceback (most recent call last): [ 594.476175] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 594.476175] env[62460]: listener.cb(fileno) [ 594.476175] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 594.476175] env[62460]: result = function(*args, **kwargs) [ 594.476175] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 594.476175] env[62460]: return func(*args, **kwargs) [ 594.476175] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 594.476175] env[62460]: raise e [ 594.476175] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 594.476175] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 594.476175] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 594.476175] env[62460]: created_port_ids = self._update_ports_for_instance( [ 594.476175] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 594.476175] env[62460]: with excutils.save_and_reraise_exception(): [ 594.476175] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 594.476175] env[62460]: self.force_reraise() [ 594.476175] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 594.476175] env[62460]: raise self.value [ 594.476175] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 594.476175] env[62460]: updated_port = self._update_port( [ 594.476175] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 594.476175] env[62460]: _ensure_no_port_binding_failure(port) [ 594.476175] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 594.476175] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 594.476779] env[62460]: nova.exception.PortBindingFailed: Binding failed for port 536f1463-299b-4c7b-8962-741db114114c, please check neutron logs for more information. [ 594.476779] env[62460]: Removing descriptor: 18 [ 594.476779] env[62460]: ERROR nova.compute.manager [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 536f1463-299b-4c7b-8962-741db114114c, please check neutron logs for more information. [ 594.476779] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] Traceback (most recent call last): [ 594.476779] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 594.476779] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] yield resources [ 594.476779] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 594.476779] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] self.driver.spawn(context, instance, image_meta, [ 594.476779] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 594.476779] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] self._vmops.spawn(context, instance, image_meta, injected_files, [ 594.476779] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 594.476779] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] vm_ref = self.build_virtual_machine(instance, [ 594.477119] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 594.477119] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] vif_infos = vmwarevif.get_vif_info(self._session, [ 594.477119] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 594.477119] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] for vif in network_info: [ 594.477119] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 594.477119] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] return self._sync_wrapper(fn, *args, **kwargs) [ 594.477119] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 594.477119] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] self.wait() [ 594.477119] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 594.477119] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] self[:] = self._gt.wait() [ 594.477119] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 594.477119] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] return self._exit_event.wait() [ 594.477119] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 594.477394] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] result = hub.switch() [ 594.477394] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 594.477394] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] return self.greenlet.switch() [ 594.477394] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 594.477394] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] result = function(*args, **kwargs) [ 594.477394] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 594.477394] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] return func(*args, **kwargs) [ 594.477394] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 594.477394] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] raise e [ 594.477394] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 594.477394] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] nwinfo = self.network_api.allocate_for_instance( [ 594.477394] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 594.477394] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] created_port_ids = self._update_ports_for_instance( [ 594.477672] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 594.477672] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] with excutils.save_and_reraise_exception(): [ 594.477672] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 594.477672] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] self.force_reraise() [ 594.477672] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 594.477672] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] raise self.value [ 594.477672] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 594.477672] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] updated_port = self._update_port( [ 594.477672] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 594.477672] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] _ensure_no_port_binding_failure(port) [ 594.477672] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 594.477672] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] raise exception.PortBindingFailed(port_id=port['id']) [ 594.477914] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] nova.exception.PortBindingFailed: Binding failed for port 536f1463-299b-4c7b-8962-741db114114c, please check neutron logs for more information. [ 594.477914] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] [ 594.477914] env[62460]: INFO nova.compute.manager [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] Terminating instance [ 594.477914] env[62460]: DEBUG oslo_concurrency.lockutils [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] Acquiring lock "refresh_cache-a8d7055e-e7a3-470a-ae96-f89753afa9df" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 594.477914] env[62460]: DEBUG oslo_concurrency.lockutils [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] Acquired lock "refresh_cache-a8d7055e-e7a3-470a-ae96-f89753afa9df" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 594.477914] env[62460]: DEBUG nova.network.neutron [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 594.812703] env[62460]: DEBUG oslo_concurrency.lockutils [req-b61e841c-3a62-4198-9354-2a7b03df7f96 req-a783c17a-5c60-4bb2-91d2-939ab68141c5 service nova] Releasing lock "refresh_cache-9b8ff88e-9e13-42b2-92b0-13af6873c916" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 594.812960] env[62460]: DEBUG nova.compute.manager [req-b61e841c-3a62-4198-9354-2a7b03df7f96 req-a783c17a-5c60-4bb2-91d2-939ab68141c5 service nova] [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] Received event network-vif-deleted-14e5de25-c06c-48c1-a690-db4605dc0193 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 594.833580] env[62460]: DEBUG nova.network.neutron [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 594.952011] env[62460]: DEBUG nova.network.neutron [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 595.004201] env[62460]: DEBUG nova.network.neutron [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 595.152340] env[62460]: DEBUG nova.network.neutron [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 595.455247] env[62460]: DEBUG oslo_concurrency.lockutils [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Releasing lock "refresh_cache-efef4327-22c8-4b59-9b26-89bcfe89bc59" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 595.455749] env[62460]: DEBUG nova.compute.manager [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 595.455873] env[62460]: DEBUG nova.compute.manager [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 595.456060] env[62460]: DEBUG nova.network.neutron [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 595.490019] env[62460]: DEBUG nova.network.neutron [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 595.597910] env[62460]: DEBUG nova.compute.manager [req-a266262e-3f9a-4828-868f-66a84e17652b req-7c877bc2-dd95-4ca5-8603-5920fc578c0e service nova] [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] Received event network-changed-536f1463-299b-4c7b-8962-741db114114c {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 595.598296] env[62460]: DEBUG nova.compute.manager [req-a266262e-3f9a-4828-868f-66a84e17652b req-7c877bc2-dd95-4ca5-8603-5920fc578c0e service nova] [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] Refreshing instance network info cache due to event network-changed-536f1463-299b-4c7b-8962-741db114114c. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 595.598802] env[62460]: DEBUG oslo_concurrency.lockutils [req-a266262e-3f9a-4828-868f-66a84e17652b req-7c877bc2-dd95-4ca5-8603-5920fc578c0e service nova] Acquiring lock "refresh_cache-a8d7055e-e7a3-470a-ae96-f89753afa9df" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 595.659019] env[62460]: DEBUG oslo_concurrency.lockutils [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] Releasing lock "refresh_cache-a8d7055e-e7a3-470a-ae96-f89753afa9df" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 595.659019] env[62460]: DEBUG nova.compute.manager [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 595.659019] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 595.659019] env[62460]: DEBUG oslo_concurrency.lockutils [req-a266262e-3f9a-4828-868f-66a84e17652b req-7c877bc2-dd95-4ca5-8603-5920fc578c0e service nova] Acquired lock "refresh_cache-a8d7055e-e7a3-470a-ae96-f89753afa9df" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 595.659019] env[62460]: DEBUG nova.network.neutron [req-a266262e-3f9a-4828-868f-66a84e17652b req-7c877bc2-dd95-4ca5-8603-5920fc578c0e service nova] [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] Refreshing network info cache for port 536f1463-299b-4c7b-8962-741db114114c {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 595.659243] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6fc02d7f-0518-46db-96c9-b72b70c54004 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.672404] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf504c6f-dbd1-4278-890c-4d7238c467e4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.696035] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] Acquiring lock "47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 595.697300] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] Lock "47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 595.710016] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a8d7055e-e7a3-470a-ae96-f89753afa9df could not be found. [ 595.710016] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 595.710016] env[62460]: INFO nova.compute.manager [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] Took 0.05 seconds to destroy the instance on the hypervisor. [ 595.710016] env[62460]: DEBUG oslo.service.loopingcall [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 595.712311] env[62460]: DEBUG nova.compute.manager [-] [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 595.712414] env[62460]: DEBUG nova.network.neutron [-] [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 595.732551] env[62460]: DEBUG nova.network.neutron [-] [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 595.790215] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df49b612-50db-4e4c-aae9-1121466064c4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.797765] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ee7cf39-c65e-4fd2-a9ab-3663d5120473 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.834742] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4ecab96-cd53-4505-b11c-b74d7f0d85e4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.843044] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f610cb1-e7a4-4c55-8c5a-e97b7800f9a1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.856297] env[62460]: DEBUG nova.compute.provider_tree [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 595.990593] env[62460]: DEBUG nova.network.neutron [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 596.186632] env[62460]: DEBUG nova.network.neutron [req-a266262e-3f9a-4828-868f-66a84e17652b req-7c877bc2-dd95-4ca5-8603-5920fc578c0e service nova] [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 596.236841] env[62460]: DEBUG nova.network.neutron [-] [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 596.334434] env[62460]: DEBUG nova.network.neutron [req-a266262e-3f9a-4828-868f-66a84e17652b req-7c877bc2-dd95-4ca5-8603-5920fc578c0e service nova] [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 596.360176] env[62460]: DEBUG nova.scheduler.client.report [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 596.495444] env[62460]: INFO nova.compute.manager [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: efef4327-22c8-4b59-9b26-89bcfe89bc59] Took 1.04 seconds to deallocate network for instance. [ 596.737347] env[62460]: INFO nova.compute.manager [-] [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] Took 1.02 seconds to deallocate network for instance. [ 596.740115] env[62460]: DEBUG nova.compute.claims [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 596.740115] env[62460]: DEBUG oslo_concurrency.lockutils [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 596.838647] env[62460]: DEBUG oslo_concurrency.lockutils [req-a266262e-3f9a-4828-868f-66a84e17652b req-7c877bc2-dd95-4ca5-8603-5920fc578c0e service nova] Releasing lock "refresh_cache-a8d7055e-e7a3-470a-ae96-f89753afa9df" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 596.838647] env[62460]: DEBUG nova.compute.manager [req-a266262e-3f9a-4828-868f-66a84e17652b req-7c877bc2-dd95-4ca5-8603-5920fc578c0e service nova] [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] Received event network-vif-deleted-536f1463-299b-4c7b-8962-741db114114c {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 596.866978] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.581s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 596.867434] env[62460]: DEBUG nova.compute.manager [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] [instance: a2747204-95c5-4200-8742-9ec39b4368c8] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 596.870754] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.477s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.376999] env[62460]: DEBUG nova.compute.utils [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 597.381045] env[62460]: DEBUG nova.compute.manager [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] [instance: a2747204-95c5-4200-8742-9ec39b4368c8] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 597.381244] env[62460]: DEBUG nova.network.neutron [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] [instance: a2747204-95c5-4200-8742-9ec39b4368c8] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 597.439919] env[62460]: DEBUG nova.policy [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd7cc445eeb544e9386048a106cf9a936', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2908d3b3f37241b8ba5ffa78d6150d48', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 597.518251] env[62460]: INFO nova.scheduler.client.report [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Deleted allocations for instance efef4327-22c8-4b59-9b26-89bcfe89bc59 [ 597.813981] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81b79d5e-cd83-46aa-9d13-f6c7a6cb0bf2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.822292] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8e65771-3baf-47db-84be-7946acafcf30 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.860854] env[62460]: DEBUG nova.network.neutron [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] [instance: a2747204-95c5-4200-8742-9ec39b4368c8] Successfully created port: 660a85f6-adeb-4e4c-be35-ae655ea19cfd {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 597.865349] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a832939e-6976-4599-98f9-671eb3c9e4a6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.873881] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9643f50b-5bfc-4eff-9798-0e21ce59bb79 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.888541] env[62460]: DEBUG nova.compute.manager [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] [instance: a2747204-95c5-4200-8742-9ec39b4368c8] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 597.891839] env[62460]: DEBUG nova.compute.provider_tree [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 597.916384] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] Acquiring lock "770b6195-ab31-4df0-84b0-e8382732ec32" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 597.916664] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] Lock "770b6195-ab31-4df0-84b0-e8382732ec32" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 598.031011] env[62460]: DEBUG oslo_concurrency.lockutils [None req-dd06b6bd-53e3-4181-b64d-c8a5970bd170 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Lock "efef4327-22c8-4b59-9b26-89bcfe89bc59" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 54.636s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 598.394572] env[62460]: INFO nova.virt.block_device [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] [instance: a2747204-95c5-4200-8742-9ec39b4368c8] Booting with volume 20188ab7-fec7-4933-8022-479219d3e31e at /dev/sda [ 598.397283] env[62460]: DEBUG nova.scheduler.client.report [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 598.450154] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8cf7fdcc-70f3-4a79-8925-88fb7156ab7a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.458922] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bc4dd8d-a511-4528-a16b-f699ad31ed19 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.480020] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a676d310-0ca0-4a1a-9f66-b0791d89040b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.487543] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e0e9b0b-a2b1-4e3b-9d5c-de8f5369493c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.508014] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d493a53d-fb83-4dfa-8eb3-7d9664707497 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.514023] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91032157-9458-40cd-95d5-25e21c5ecae6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.526917] env[62460]: DEBUG nova.virt.block_device [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] [instance: a2747204-95c5-4200-8742-9ec39b4368c8] Updating existing volume attachment record: c55eeaef-2de8-4813-8d46-ba7eafda95b8 {{(pid=62460) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 598.537356] env[62460]: DEBUG nova.compute.manager [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] [instance: 3b71c366-cbd9-4b98-aa0d-c55b56e69231] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 598.759335] env[62460]: DEBUG nova.compute.manager [req-f401eb31-e060-48d7-a7b7-153948785aa2 req-8d3d652b-ae38-4719-9e9b-e895640c9663 service nova] [instance: a2747204-95c5-4200-8742-9ec39b4368c8] Received event network-changed-660a85f6-adeb-4e4c-be35-ae655ea19cfd {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 598.759558] env[62460]: DEBUG nova.compute.manager [req-f401eb31-e060-48d7-a7b7-153948785aa2 req-8d3d652b-ae38-4719-9e9b-e895640c9663 service nova] [instance: a2747204-95c5-4200-8742-9ec39b4368c8] Refreshing instance network info cache due to event network-changed-660a85f6-adeb-4e4c-be35-ae655ea19cfd. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 598.759770] env[62460]: DEBUG oslo_concurrency.lockutils [req-f401eb31-e060-48d7-a7b7-153948785aa2 req-8d3d652b-ae38-4719-9e9b-e895640c9663 service nova] Acquiring lock "refresh_cache-a2747204-95c5-4200-8742-9ec39b4368c8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 598.759883] env[62460]: DEBUG oslo_concurrency.lockutils [req-f401eb31-e060-48d7-a7b7-153948785aa2 req-8d3d652b-ae38-4719-9e9b-e895640c9663 service nova] Acquired lock "refresh_cache-a2747204-95c5-4200-8742-9ec39b4368c8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 598.760064] env[62460]: DEBUG nova.network.neutron [req-f401eb31-e060-48d7-a7b7-153948785aa2 req-8d3d652b-ae38-4719-9e9b-e895640c9663 service nova] [instance: a2747204-95c5-4200-8742-9ec39b4368c8] Refreshing network info cache for port 660a85f6-adeb-4e4c-be35-ae655ea19cfd {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 598.916319] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.046s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 598.917947] env[62460]: ERROR nova.compute.manager [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0978e846-4023-4bcd-867b-55313312ab35, please check neutron logs for more information. [ 598.917947] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] Traceback (most recent call last): [ 598.917947] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 598.917947] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] self.driver.spawn(context, instance, image_meta, [ 598.917947] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 598.917947] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 598.917947] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 598.917947] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] vm_ref = self.build_virtual_machine(instance, [ 598.917947] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 598.917947] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] vif_infos = vmwarevif.get_vif_info(self._session, [ 598.917947] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 598.918248] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] for vif in network_info: [ 598.918248] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 598.918248] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] return self._sync_wrapper(fn, *args, **kwargs) [ 598.918248] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 598.918248] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] self.wait() [ 598.918248] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 598.918248] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] self[:] = self._gt.wait() [ 598.918248] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 598.918248] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] return self._exit_event.wait() [ 598.918248] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 598.918248] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] result = hub.switch() [ 598.918248] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 598.918248] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] return self.greenlet.switch() [ 598.918516] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 598.918516] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] result = function(*args, **kwargs) [ 598.918516] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 598.918516] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] return func(*args, **kwargs) [ 598.918516] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 598.918516] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] raise e [ 598.918516] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 598.918516] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] nwinfo = self.network_api.allocate_for_instance( [ 598.918516] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 598.918516] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] created_port_ids = self._update_ports_for_instance( [ 598.918516] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 598.918516] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] with excutils.save_and_reraise_exception(): [ 598.918516] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 598.918784] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] self.force_reraise() [ 598.918784] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 598.918784] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] raise self.value [ 598.918784] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 598.918784] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] updated_port = self._update_port( [ 598.918784] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 598.918784] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] _ensure_no_port_binding_failure(port) [ 598.918784] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 598.918784] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] raise exception.PortBindingFailed(port_id=port['id']) [ 598.918784] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] nova.exception.PortBindingFailed: Binding failed for port 0978e846-4023-4bcd-867b-55313312ab35, please check neutron logs for more information. [ 598.918784] env[62460]: ERROR nova.compute.manager [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] [ 598.922231] env[62460]: DEBUG nova.compute.utils [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] Binding failed for port 0978e846-4023-4bcd-867b-55313312ab35, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 598.922231] env[62460]: DEBUG oslo_concurrency.lockutils [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.395s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 598.925382] env[62460]: DEBUG nova.compute.manager [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] Build of instance a3f7edbe-4bba-4fff-9e62-99b7a85f971b was re-scheduled: Binding failed for port 0978e846-4023-4bcd-867b-55313312ab35, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 598.925842] env[62460]: DEBUG nova.compute.manager [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 598.926505] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] Acquiring lock "refresh_cache-a3f7edbe-4bba-4fff-9e62-99b7a85f971b" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 598.926505] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] Acquired lock "refresh_cache-a3f7edbe-4bba-4fff-9e62-99b7a85f971b" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 598.926604] env[62460]: DEBUG nova.network.neutron [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 598.999446] env[62460]: ERROR nova.compute.manager [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 660a85f6-adeb-4e4c-be35-ae655ea19cfd, please check neutron logs for more information. [ 598.999446] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 598.999446] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 598.999446] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 598.999446] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 598.999446] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 598.999446] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 598.999446] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 598.999446] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 598.999446] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 598.999446] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 598.999446] env[62460]: ERROR nova.compute.manager raise self.value [ 598.999446] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 598.999446] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 598.999446] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 598.999446] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 598.999792] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 598.999792] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 598.999792] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 660a85f6-adeb-4e4c-be35-ae655ea19cfd, please check neutron logs for more information. [ 598.999792] env[62460]: ERROR nova.compute.manager [ 598.999792] env[62460]: Traceback (most recent call last): [ 598.999792] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 598.999792] env[62460]: listener.cb(fileno) [ 598.999792] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 598.999792] env[62460]: result = function(*args, **kwargs) [ 598.999792] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 598.999792] env[62460]: return func(*args, **kwargs) [ 598.999792] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 598.999792] env[62460]: raise e [ 598.999792] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 598.999792] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 598.999792] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 598.999792] env[62460]: created_port_ids = self._update_ports_for_instance( [ 598.999792] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 598.999792] env[62460]: with excutils.save_and_reraise_exception(): [ 598.999792] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 598.999792] env[62460]: self.force_reraise() [ 598.999792] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 598.999792] env[62460]: raise self.value [ 598.999792] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 598.999792] env[62460]: updated_port = self._update_port( [ 598.999792] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 598.999792] env[62460]: _ensure_no_port_binding_failure(port) [ 598.999792] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 598.999792] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 599.000520] env[62460]: nova.exception.PortBindingFailed: Binding failed for port 660a85f6-adeb-4e4c-be35-ae655ea19cfd, please check neutron logs for more information. [ 599.000520] env[62460]: Removing descriptor: 18 [ 599.066488] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 599.277895] env[62460]: DEBUG nova.network.neutron [req-f401eb31-e060-48d7-a7b7-153948785aa2 req-8d3d652b-ae38-4719-9e9b-e895640c9663 service nova] [instance: a2747204-95c5-4200-8742-9ec39b4368c8] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 599.363052] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 599.363141] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 599.367536] env[62460]: DEBUG nova.network.neutron [req-f401eb31-e060-48d7-a7b7-153948785aa2 req-8d3d652b-ae38-4719-9e9b-e895640c9663 service nova] [instance: a2747204-95c5-4200-8742-9ec39b4368c8] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 599.449895] env[62460]: DEBUG nova.network.neutron [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 599.542029] env[62460]: DEBUG nova.network.neutron [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 599.803729] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-807a2d05-4d32-4e6e-9eca-f0044ea07956 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.811155] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7249a4fb-591a-43a4-a972-3b1b23b9851b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.841100] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68f55915-8aca-43e0-acb2-dae9a1e20fba {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.848565] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79172321-c2e9-4ac3-808d-7564cd1aa9a0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.862857] env[62460]: DEBUG nova.compute.provider_tree [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 599.868471] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 599.868546] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Starting heal instance info cache {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 599.868774] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Rebuilding the list of instances to heal {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 599.869888] env[62460]: DEBUG oslo_concurrency.lockutils [req-f401eb31-e060-48d7-a7b7-153948785aa2 req-8d3d652b-ae38-4719-9e9b-e895640c9663 service nova] Releasing lock "refresh_cache-a2747204-95c5-4200-8742-9ec39b4368c8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 600.044399] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] Releasing lock "refresh_cache-a3f7edbe-4bba-4fff-9e62-99b7a85f971b" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 600.045169] env[62460]: DEBUG nova.compute.manager [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 600.045169] env[62460]: DEBUG nova.compute.manager [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 600.045169] env[62460]: DEBUG nova.network.neutron [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 600.062304] env[62460]: DEBUG nova.network.neutron [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 600.366743] env[62460]: DEBUG nova.scheduler.client.report [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 600.376017] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] Skipping network cache update for instance because it is Building. {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 600.376017] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] Skipping network cache update for instance because it is Building. {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 600.376017] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] Skipping network cache update for instance because it is Building. {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 600.376017] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] Skipping network cache update for instance because it is Building. {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 600.376017] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] Skipping network cache update for instance because it is Building. {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 600.376017] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: a2747204-95c5-4200-8742-9ec39b4368c8] Skipping network cache update for instance because it is Building. {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 600.390218] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Acquiring lock "refresh_cache-00544d2a-1a15-4347-abe4-3641b5bf0cfe" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 600.390557] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Acquired lock "refresh_cache-00544d2a-1a15-4347-abe4-3641b5bf0cfe" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 600.390557] env[62460]: DEBUG nova.network.neutron [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 00544d2a-1a15-4347-abe4-3641b5bf0cfe] Forcefully refreshing network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 600.390663] env[62460]: DEBUG nova.objects.instance [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lazy-loading 'info_cache' on Instance uuid 00544d2a-1a15-4347-abe4-3641b5bf0cfe {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 600.565227] env[62460]: DEBUG nova.network.neutron [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 600.638975] env[62460]: DEBUG nova.compute.manager [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] [instance: a2747204-95c5-4200-8742-9ec39b4368c8] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 600.638975] env[62460]: DEBUG nova.virt.hardware [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 600.638975] env[62460]: DEBUG nova.virt.hardware [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 600.639253] env[62460]: DEBUG nova.virt.hardware [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 600.639482] env[62460]: DEBUG nova.virt.hardware [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 600.639674] env[62460]: DEBUG nova.virt.hardware [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 600.639850] env[62460]: DEBUG nova.virt.hardware [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 600.640102] env[62460]: DEBUG nova.virt.hardware [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 600.640306] env[62460]: DEBUG nova.virt.hardware [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 600.640509] env[62460]: DEBUG nova.virt.hardware [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 600.640704] env[62460]: DEBUG nova.virt.hardware [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 600.640923] env[62460]: DEBUG nova.virt.hardware [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 600.642205] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a77df89-f800-4354-8297-08fa0de7bb94 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.650792] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65264ddc-b212-4d3b-be60-161e84951b57 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.665314] env[62460]: ERROR nova.compute.manager [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] [instance: a2747204-95c5-4200-8742-9ec39b4368c8] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 660a85f6-adeb-4e4c-be35-ae655ea19cfd, please check neutron logs for more information. [ 600.665314] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] Traceback (most recent call last): [ 600.665314] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 600.665314] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] yield resources [ 600.665314] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 600.665314] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] self.driver.spawn(context, instance, image_meta, [ 600.665314] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 600.665314] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 600.665314] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 600.665314] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] vm_ref = self.build_virtual_machine(instance, [ 600.665314] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 600.665663] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] vif_infos = vmwarevif.get_vif_info(self._session, [ 600.665663] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 600.665663] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] for vif in network_info: [ 600.665663] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 600.665663] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] return self._sync_wrapper(fn, *args, **kwargs) [ 600.665663] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 600.665663] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] self.wait() [ 600.665663] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 600.665663] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] self[:] = self._gt.wait() [ 600.665663] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 600.665663] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] return self._exit_event.wait() [ 600.665663] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 600.665663] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] current.throw(*self._exc) [ 600.665998] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 600.665998] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] result = function(*args, **kwargs) [ 600.665998] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 600.665998] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] return func(*args, **kwargs) [ 600.665998] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 600.665998] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] raise e [ 600.665998] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 600.665998] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] nwinfo = self.network_api.allocate_for_instance( [ 600.665998] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 600.665998] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] created_port_ids = self._update_ports_for_instance( [ 600.665998] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 600.665998] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] with excutils.save_and_reraise_exception(): [ 600.665998] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 600.666351] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] self.force_reraise() [ 600.666351] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 600.666351] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] raise self.value [ 600.666351] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 600.666351] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] updated_port = self._update_port( [ 600.666351] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 600.666351] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] _ensure_no_port_binding_failure(port) [ 600.666351] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 600.666351] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] raise exception.PortBindingFailed(port_id=port['id']) [ 600.666351] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] nova.exception.PortBindingFailed: Binding failed for port 660a85f6-adeb-4e4c-be35-ae655ea19cfd, please check neutron logs for more information. [ 600.666351] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] [ 600.666351] env[62460]: INFO nova.compute.manager [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] [instance: a2747204-95c5-4200-8742-9ec39b4368c8] Terminating instance [ 600.667684] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] Acquiring lock "refresh_cache-a2747204-95c5-4200-8742-9ec39b4368c8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 600.667872] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] Acquired lock "refresh_cache-a2747204-95c5-4200-8742-9ec39b4368c8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 600.668060] env[62460]: DEBUG nova.network.neutron [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] [instance: a2747204-95c5-4200-8742-9ec39b4368c8] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 600.806967] env[62460]: DEBUG nova.compute.manager [req-ec7bad9e-87f4-4f3b-a060-71aa7503898a req-ddd2bb15-f3e1-4c56-b9f4-259e646e1886 service nova] [instance: a2747204-95c5-4200-8742-9ec39b4368c8] Received event network-vif-deleted-660a85f6-adeb-4e4c-be35-ae655ea19cfd {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 600.879195] env[62460]: DEBUG oslo_concurrency.lockutils [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.957s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 600.879195] env[62460]: ERROR nova.compute.manager [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 53ec4989-7198-480a-ab0a-351544564777, please check neutron logs for more information. [ 600.879195] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] Traceback (most recent call last): [ 600.879195] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 600.879195] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] self.driver.spawn(context, instance, image_meta, [ 600.879195] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 600.879195] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 600.879195] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 600.879195] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] vm_ref = self.build_virtual_machine(instance, [ 600.879465] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 600.879465] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] vif_infos = vmwarevif.get_vif_info(self._session, [ 600.879465] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 600.879465] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] for vif in network_info: [ 600.879465] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 600.879465] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] return self._sync_wrapper(fn, *args, **kwargs) [ 600.879465] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 600.879465] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] self.wait() [ 600.879465] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 600.879465] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] self[:] = self._gt.wait() [ 600.879465] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 600.879465] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] return self._exit_event.wait() [ 600.879465] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 600.879748] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] result = hub.switch() [ 600.879748] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 600.879748] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] return self.greenlet.switch() [ 600.879748] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 600.879748] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] result = function(*args, **kwargs) [ 600.879748] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 600.879748] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] return func(*args, **kwargs) [ 600.879748] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 600.879748] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] raise e [ 600.879748] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 600.879748] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] nwinfo = self.network_api.allocate_for_instance( [ 600.879748] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 600.879748] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] created_port_ids = self._update_ports_for_instance( [ 600.880112] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 600.880112] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] with excutils.save_and_reraise_exception(): [ 600.880112] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 600.880112] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] self.force_reraise() [ 600.880112] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 600.880112] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] raise self.value [ 600.880112] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 600.880112] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] updated_port = self._update_port( [ 600.880112] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 600.880112] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] _ensure_no_port_binding_failure(port) [ 600.880112] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 600.880112] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] raise exception.PortBindingFailed(port_id=port['id']) [ 600.880367] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] nova.exception.PortBindingFailed: Binding failed for port 53ec4989-7198-480a-ab0a-351544564777, please check neutron logs for more information. [ 600.880367] env[62460]: ERROR nova.compute.manager [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] [ 600.880367] env[62460]: DEBUG nova.compute.utils [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] Binding failed for port 53ec4989-7198-480a-ab0a-351544564777, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 600.881037] env[62460]: DEBUG oslo_concurrency.lockutils [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.215s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 600.882327] env[62460]: INFO nova.compute.claims [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 600.885891] env[62460]: DEBUG nova.compute.manager [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] Build of instance 8a2699f3-3cc4-4ec3-899a-216e09ad086a was re-scheduled: Binding failed for port 53ec4989-7198-480a-ab0a-351544564777, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 600.886525] env[62460]: DEBUG nova.compute.manager [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 600.886700] env[62460]: DEBUG oslo_concurrency.lockutils [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Acquiring lock "refresh_cache-8a2699f3-3cc4-4ec3-899a-216e09ad086a" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 600.886865] env[62460]: DEBUG oslo_concurrency.lockutils [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Acquired lock "refresh_cache-8a2699f3-3cc4-4ec3-899a-216e09ad086a" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 600.887037] env[62460]: DEBUG nova.network.neutron [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 601.068030] env[62460]: INFO nova.compute.manager [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] [instance: a3f7edbe-4bba-4fff-9e62-99b7a85f971b] Took 1.02 seconds to deallocate network for instance. [ 601.185857] env[62460]: DEBUG nova.network.neutron [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] [instance: a2747204-95c5-4200-8742-9ec39b4368c8] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 601.277024] env[62460]: DEBUG nova.network.neutron [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] [instance: a2747204-95c5-4200-8742-9ec39b4368c8] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 601.409438] env[62460]: DEBUG nova.network.neutron [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 601.420663] env[62460]: DEBUG nova.network.neutron [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 00544d2a-1a15-4347-abe4-3641b5bf0cfe] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 601.457769] env[62460]: DEBUG nova.network.neutron [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 601.780460] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] Releasing lock "refresh_cache-a2747204-95c5-4200-8742-9ec39b4368c8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 601.781080] env[62460]: DEBUG nova.compute.manager [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] [instance: a2747204-95c5-4200-8742-9ec39b4368c8] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 601.781426] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ac0e2b66-1a75-49fc-955f-d44edfac4b27 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.790571] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bd4c1c3-23d2-4cab-b066-615b499e0f38 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.812454] env[62460]: WARNING nova.virt.vmwareapi.driver [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] [instance: a2747204-95c5-4200-8742-9ec39b4368c8] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance a2747204-95c5-4200-8742-9ec39b4368c8 could not be found. [ 601.812676] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] [instance: a2747204-95c5-4200-8742-9ec39b4368c8] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 601.812942] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7627b0d3-1ecc-4bb3-a3e8-4c01a6bc05e4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.820386] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23c542dc-15bf-42c7-a9f7-cbb1c7631ad6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.841583] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] [instance: a2747204-95c5-4200-8742-9ec39b4368c8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a2747204-95c5-4200-8742-9ec39b4368c8 could not be found. [ 601.841583] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] [instance: a2747204-95c5-4200-8742-9ec39b4368c8] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 601.841583] env[62460]: INFO nova.compute.manager [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] [instance: a2747204-95c5-4200-8742-9ec39b4368c8] Took 0.06 seconds to destroy the instance on the hypervisor. [ 601.841829] env[62460]: DEBUG oslo.service.loopingcall [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 601.842047] env[62460]: DEBUG nova.compute.manager [-] [instance: a2747204-95c5-4200-8742-9ec39b4368c8] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 601.842166] env[62460]: DEBUG nova.network.neutron [-] [instance: a2747204-95c5-4200-8742-9ec39b4368c8] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 601.859874] env[62460]: DEBUG nova.network.neutron [-] [instance: a2747204-95c5-4200-8742-9ec39b4368c8] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 601.962421] env[62460]: DEBUG oslo_concurrency.lockutils [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Releasing lock "refresh_cache-8a2699f3-3cc4-4ec3-899a-216e09ad086a" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 601.962661] env[62460]: DEBUG nova.compute.manager [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 601.962828] env[62460]: DEBUG nova.compute.manager [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 601.963031] env[62460]: DEBUG nova.network.neutron [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 601.977744] env[62460]: DEBUG nova.network.neutron [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 601.988755] env[62460]: DEBUG nova.network.neutron [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 00544d2a-1a15-4347-abe4-3641b5bf0cfe] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.097450] env[62460]: INFO nova.scheduler.client.report [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] Deleted allocations for instance a3f7edbe-4bba-4fff-9e62-99b7a85f971b [ 602.293774] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fe79580-ce4d-4b9a-9d45-fbb68b7a33fc {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.302902] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2b95ea1-7129-4a92-b665-8284523b14e8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.333440] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad17fbe4-4b9a-41c0-a464-369795aeead3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.340660] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daa6a93e-603a-4583-8995-530ee6d64d63 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.353478] env[62460]: DEBUG nova.compute.provider_tree [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 602.362497] env[62460]: DEBUG nova.network.neutron [-] [instance: a2747204-95c5-4200-8742-9ec39b4368c8] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.482142] env[62460]: DEBUG nova.network.neutron [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.491539] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Releasing lock "refresh_cache-00544d2a-1a15-4347-abe4-3641b5bf0cfe" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 602.492815] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 00544d2a-1a15-4347-abe4-3641b5bf0cfe] Updated the network info_cache for instance {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 602.493871] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 602.493871] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 602.493871] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 602.493871] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 602.493871] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 602.494069] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 602.494161] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62460) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 602.494909] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 602.613486] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ffe7749c-d4a4-4eda-9bc8-8e2391dcfb20 tempest-FloatingIPsAssociationNegativeTestJSON-506448877 tempest-FloatingIPsAssociationNegativeTestJSON-506448877-project-member] Lock "a3f7edbe-4bba-4fff-9e62-99b7a85f971b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 55.502s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 602.858506] env[62460]: DEBUG nova.scheduler.client.report [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 602.865894] env[62460]: INFO nova.compute.manager [-] [instance: a2747204-95c5-4200-8742-9ec39b4368c8] Took 1.02 seconds to deallocate network for instance. [ 602.987804] env[62460]: INFO nova.compute.manager [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] [instance: 8a2699f3-3cc4-4ec3-899a-216e09ad086a] Took 1.02 seconds to deallocate network for instance. [ 602.996955] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.117173] env[62460]: DEBUG nova.compute.manager [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 603.363150] env[62460]: DEBUG oslo_concurrency.lockutils [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.482s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 603.363689] env[62460]: DEBUG nova.compute.manager [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 603.366144] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.047s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.032216] env[62460]: INFO nova.compute.manager [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] [instance: a2747204-95c5-4200-8742-9ec39b4368c8] Took 1.17 seconds to detach 1 volumes for instance. [ 604.041035] env[62460]: DEBUG nova.compute.utils [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 604.045354] env[62460]: DEBUG nova.compute.claims [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] [instance: a2747204-95c5-4200-8742-9ec39b4368c8] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 604.045444] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.047695] env[62460]: DEBUG nova.compute.manager [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 604.048113] env[62460]: DEBUG nova.network.neutron [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 604.078895] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.156035] env[62460]: DEBUG nova.policy [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a3c29d0a06624ad98155128ead12239f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '98f7cf2e2bc9462498b3c774b901136e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 604.467994] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d280cfe-eb14-4cd0-aa5e-af5329b61b71 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.476187] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc2d4876-fc3b-429b-a8ee-5c0169c7e2c2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.507301] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9acb56df-b095-4d47-a5b7-0d571b566912 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.513837] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35b2de9c-1d1e-4430-b9a4-743ba444e2ca {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.526991] env[62460]: DEBUG nova.compute.provider_tree [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 604.548822] env[62460]: DEBUG nova.compute.manager [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 604.573744] env[62460]: INFO nova.scheduler.client.report [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Deleted allocations for instance 8a2699f3-3cc4-4ec3-899a-216e09ad086a [ 604.907720] env[62460]: DEBUG nova.network.neutron [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] Successfully created port: 63267a51-1e3d-4d36-ae79-2d9931707395 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 605.030016] env[62460]: DEBUG nova.scheduler.client.report [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 605.083838] env[62460]: DEBUG oslo_concurrency.lockutils [None req-05e11c1c-1e9a-4bcb-8a1e-7d3da0e123c9 tempest-ListImageFiltersTestJSON-905139741 tempest-ListImageFiltersTestJSON-905139741-project-member] Lock "8a2699f3-3cc4-4ec3-899a-216e09ad086a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 58.911s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 605.536173] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.170s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 605.536476] env[62460]: ERROR nova.compute.manager [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e61d2785-3f7f-42cd-81ba-b8dde9aa7863, please check neutron logs for more information. [ 605.536476] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] Traceback (most recent call last): [ 605.536476] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 605.536476] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] self.driver.spawn(context, instance, image_meta, [ 605.536476] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 605.536476] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] self._vmops.spawn(context, instance, image_meta, injected_files, [ 605.536476] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 605.536476] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] vm_ref = self.build_virtual_machine(instance, [ 605.536476] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 605.536476] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] vif_infos = vmwarevif.get_vif_info(self._session, [ 605.536476] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 605.537662] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] for vif in network_info: [ 605.537662] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 605.537662] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] return self._sync_wrapper(fn, *args, **kwargs) [ 605.537662] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 605.537662] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] self.wait() [ 605.537662] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 605.537662] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] self[:] = self._gt.wait() [ 605.537662] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 605.537662] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] return self._exit_event.wait() [ 605.537662] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 605.537662] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] result = hub.switch() [ 605.537662] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 605.537662] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] return self.greenlet.switch() [ 605.537946] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 605.537946] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] result = function(*args, **kwargs) [ 605.537946] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 605.537946] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] return func(*args, **kwargs) [ 605.537946] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 605.537946] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] raise e [ 605.537946] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 605.537946] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] nwinfo = self.network_api.allocate_for_instance( [ 605.537946] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 605.537946] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] created_port_ids = self._update_ports_for_instance( [ 605.537946] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 605.537946] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] with excutils.save_and_reraise_exception(): [ 605.537946] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 605.538245] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] self.force_reraise() [ 605.538245] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 605.538245] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] raise self.value [ 605.538245] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 605.538245] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] updated_port = self._update_port( [ 605.538245] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 605.538245] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] _ensure_no_port_binding_failure(port) [ 605.538245] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 605.538245] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] raise exception.PortBindingFailed(port_id=port['id']) [ 605.538245] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] nova.exception.PortBindingFailed: Binding failed for port e61d2785-3f7f-42cd-81ba-b8dde9aa7863, please check neutron logs for more information. [ 605.538245] env[62460]: ERROR nova.compute.manager [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] [ 605.538547] env[62460]: DEBUG nova.compute.utils [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] Binding failed for port e61d2785-3f7f-42cd-81ba-b8dde9aa7863, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 605.538547] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5a702432-05a2-486c-ae26-233347bf963c tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.763s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 605.540489] env[62460]: DEBUG nova.objects.instance [None req-5a702432-05a2-486c-ae26-233347bf963c tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Lazy-loading 'resources' on Instance uuid 00544d2a-1a15-4347-abe4-3641b5bf0cfe {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 605.541924] env[62460]: DEBUG nova.compute.manager [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] Build of instance b51ddc3a-de97-4d51-adc9-e92b4fb0a150 was re-scheduled: Binding failed for port e61d2785-3f7f-42cd-81ba-b8dde9aa7863, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 605.542191] env[62460]: DEBUG nova.compute.manager [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 605.542343] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "refresh_cache-b51ddc3a-de97-4d51-adc9-e92b4fb0a150" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 605.542491] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquired lock "refresh_cache-b51ddc3a-de97-4d51-adc9-e92b4fb0a150" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 605.542649] env[62460]: DEBUG nova.network.neutron [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 605.557370] env[62460]: DEBUG nova.compute.manager [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 605.587623] env[62460]: DEBUG nova.virt.hardware [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 605.587868] env[62460]: DEBUG nova.virt.hardware [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 605.588068] env[62460]: DEBUG nova.virt.hardware [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 605.588281] env[62460]: DEBUG nova.virt.hardware [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 605.588430] env[62460]: DEBUG nova.virt.hardware [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 605.588576] env[62460]: DEBUG nova.virt.hardware [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 605.588779] env[62460]: DEBUG nova.virt.hardware [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 605.588935] env[62460]: DEBUG nova.virt.hardware [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 605.589116] env[62460]: DEBUG nova.virt.hardware [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 605.589278] env[62460]: DEBUG nova.virt.hardware [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 605.589448] env[62460]: DEBUG nova.virt.hardware [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 605.589865] env[62460]: DEBUG nova.compute.manager [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 605.593147] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2fe65ab-ac1c-4455-8b86-6e430bb28c16 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.601255] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49fb4a3c-c3e8-4d74-b706-78281ca72e49 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.880808] env[62460]: DEBUG nova.compute.manager [req-8b006127-c0ec-425b-a935-607cfac66f1c req-6bfc28c0-9345-47d1-9fe0-7a3703c61161 service nova] [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] Received event network-changed-63267a51-1e3d-4d36-ae79-2d9931707395 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 605.881081] env[62460]: DEBUG nova.compute.manager [req-8b006127-c0ec-425b-a935-607cfac66f1c req-6bfc28c0-9345-47d1-9fe0-7a3703c61161 service nova] [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] Refreshing instance network info cache due to event network-changed-63267a51-1e3d-4d36-ae79-2d9931707395. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 605.881381] env[62460]: DEBUG oslo_concurrency.lockutils [req-8b006127-c0ec-425b-a935-607cfac66f1c req-6bfc28c0-9345-47d1-9fe0-7a3703c61161 service nova] Acquiring lock "refresh_cache-443e235d-32b2-4af7-bdae-3f30e8196ee3" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 605.881526] env[62460]: DEBUG oslo_concurrency.lockutils [req-8b006127-c0ec-425b-a935-607cfac66f1c req-6bfc28c0-9345-47d1-9fe0-7a3703c61161 service nova] Acquired lock "refresh_cache-443e235d-32b2-4af7-bdae-3f30e8196ee3" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 605.881685] env[62460]: DEBUG nova.network.neutron [req-8b006127-c0ec-425b-a935-607cfac66f1c req-6bfc28c0-9345-47d1-9fe0-7a3703c61161 service nova] [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] Refreshing network info cache for port 63267a51-1e3d-4d36-ae79-2d9931707395 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 606.016154] env[62460]: ERROR nova.compute.manager [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 63267a51-1e3d-4d36-ae79-2d9931707395, please check neutron logs for more information. [ 606.016154] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 606.016154] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 606.016154] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 606.016154] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 606.016154] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 606.016154] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 606.016154] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 606.016154] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 606.016154] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 606.016154] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 606.016154] env[62460]: ERROR nova.compute.manager raise self.value [ 606.016154] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 606.016154] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 606.016154] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 606.016154] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 606.016711] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 606.016711] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 606.016711] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 63267a51-1e3d-4d36-ae79-2d9931707395, please check neutron logs for more information. [ 606.016711] env[62460]: ERROR nova.compute.manager [ 606.016711] env[62460]: Traceback (most recent call last): [ 606.016711] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 606.016711] env[62460]: listener.cb(fileno) [ 606.016711] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 606.016711] env[62460]: result = function(*args, **kwargs) [ 606.016711] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 606.016711] env[62460]: return func(*args, **kwargs) [ 606.016711] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 606.016711] env[62460]: raise e [ 606.016711] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 606.016711] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 606.016711] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 606.016711] env[62460]: created_port_ids = self._update_ports_for_instance( [ 606.016711] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 606.016711] env[62460]: with excutils.save_and_reraise_exception(): [ 606.016711] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 606.016711] env[62460]: self.force_reraise() [ 606.016711] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 606.016711] env[62460]: raise self.value [ 606.016711] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 606.016711] env[62460]: updated_port = self._update_port( [ 606.016711] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 606.016711] env[62460]: _ensure_no_port_binding_failure(port) [ 606.016711] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 606.016711] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 606.017489] env[62460]: nova.exception.PortBindingFailed: Binding failed for port 63267a51-1e3d-4d36-ae79-2d9931707395, please check neutron logs for more information. [ 606.017489] env[62460]: Removing descriptor: 18 [ 606.017489] env[62460]: ERROR nova.compute.manager [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 63267a51-1e3d-4d36-ae79-2d9931707395, please check neutron logs for more information. [ 606.017489] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] Traceback (most recent call last): [ 606.017489] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 606.017489] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] yield resources [ 606.017489] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 606.017489] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] self.driver.spawn(context, instance, image_meta, [ 606.017489] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 606.017489] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 606.017489] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 606.017489] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] vm_ref = self.build_virtual_machine(instance, [ 606.017799] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 606.017799] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] vif_infos = vmwarevif.get_vif_info(self._session, [ 606.017799] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 606.017799] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] for vif in network_info: [ 606.017799] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 606.017799] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] return self._sync_wrapper(fn, *args, **kwargs) [ 606.017799] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 606.017799] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] self.wait() [ 606.017799] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 606.017799] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] self[:] = self._gt.wait() [ 606.017799] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 606.017799] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] return self._exit_event.wait() [ 606.017799] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 606.018149] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] result = hub.switch() [ 606.018149] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 606.018149] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] return self.greenlet.switch() [ 606.018149] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 606.018149] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] result = function(*args, **kwargs) [ 606.018149] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 606.018149] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] return func(*args, **kwargs) [ 606.018149] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 606.018149] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] raise e [ 606.018149] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 606.018149] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] nwinfo = self.network_api.allocate_for_instance( [ 606.018149] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 606.018149] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] created_port_ids = self._update_ports_for_instance( [ 606.018479] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 606.018479] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] with excutils.save_and_reraise_exception(): [ 606.018479] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 606.018479] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] self.force_reraise() [ 606.018479] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 606.018479] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] raise self.value [ 606.018479] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 606.018479] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] updated_port = self._update_port( [ 606.018479] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 606.018479] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] _ensure_no_port_binding_failure(port) [ 606.018479] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 606.018479] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] raise exception.PortBindingFailed(port_id=port['id']) [ 606.018858] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] nova.exception.PortBindingFailed: Binding failed for port 63267a51-1e3d-4d36-ae79-2d9931707395, please check neutron logs for more information. [ 606.018858] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] [ 606.018858] env[62460]: INFO nova.compute.manager [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] Terminating instance [ 606.020404] env[62460]: DEBUG oslo_concurrency.lockutils [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Acquiring lock "refresh_cache-443e235d-32b2-4af7-bdae-3f30e8196ee3" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 606.074860] env[62460]: DEBUG nova.network.neutron [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 606.120959] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.248047] env[62460]: DEBUG nova.network.neutron [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.414322] env[62460]: DEBUG nova.network.neutron [req-8b006127-c0ec-425b-a935-607cfac66f1c req-6bfc28c0-9345-47d1-9fe0-7a3703c61161 service nova] [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 606.433595] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40011966-70e2-457a-a511-d595c31fd980 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.442435] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81330255-f0ee-4bc0-8095-55283c017f19 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.472240] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-524c8d52-424f-4e6d-ba7d-721bf68c46ec {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.479569] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-518b0dbb-ca74-4dc0-9a68-3b1f8e625733 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.492732] env[62460]: DEBUG nova.compute.provider_tree [None req-5a702432-05a2-486c-ae26-233347bf963c tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 606.504970] env[62460]: DEBUG nova.network.neutron [req-8b006127-c0ec-425b-a935-607cfac66f1c req-6bfc28c0-9345-47d1-9fe0-7a3703c61161 service nova] [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.753961] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Releasing lock "refresh_cache-b51ddc3a-de97-4d51-adc9-e92b4fb0a150" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 606.754385] env[62460]: DEBUG nova.compute.manager [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 606.755289] env[62460]: DEBUG nova.compute.manager [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 606.757031] env[62460]: DEBUG nova.network.neutron [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 606.779553] env[62460]: DEBUG nova.network.neutron [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 606.996283] env[62460]: DEBUG nova.scheduler.client.report [None req-5a702432-05a2-486c-ae26-233347bf963c tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 607.007758] env[62460]: DEBUG oslo_concurrency.lockutils [req-8b006127-c0ec-425b-a935-607cfac66f1c req-6bfc28c0-9345-47d1-9fe0-7a3703c61161 service nova] Releasing lock "refresh_cache-443e235d-32b2-4af7-bdae-3f30e8196ee3" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 607.008536] env[62460]: DEBUG oslo_concurrency.lockutils [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Acquired lock "refresh_cache-443e235d-32b2-4af7-bdae-3f30e8196ee3" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 607.008536] env[62460]: DEBUG nova.network.neutron [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 607.282370] env[62460]: DEBUG nova.network.neutron [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.508549] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5a702432-05a2-486c-ae26-233347bf963c tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.970s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 607.515126] env[62460]: DEBUG oslo_concurrency.lockutils [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.832s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 607.535265] env[62460]: DEBUG nova.network.neutron [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 607.547262] env[62460]: INFO nova.scheduler.client.report [None req-5a702432-05a2-486c-ae26-233347bf963c tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Deleted allocations for instance 00544d2a-1a15-4347-abe4-3641b5bf0cfe [ 607.643063] env[62460]: DEBUG nova.network.neutron [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.788172] env[62460]: INFO nova.compute.manager [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: b51ddc3a-de97-4d51-adc9-e92b4fb0a150] Took 1.03 seconds to deallocate network for instance. [ 608.060660] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5a702432-05a2-486c-ae26-233347bf963c tempest-ServerDiagnosticsV248Test-2057442527 tempest-ServerDiagnosticsV248Test-2057442527-project-member] Lock "00544d2a-1a15-4347-abe4-3641b5bf0cfe" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.201s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 608.089424] env[62460]: DEBUG nova.compute.manager [req-8b7556ec-e56d-4cd1-8c91-6c13d91d9913 req-c5518b76-e109-4701-8388-05917e2c341b service nova] [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] Received event network-vif-deleted-63267a51-1e3d-4d36-ae79-2d9931707395 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 608.146545] env[62460]: DEBUG oslo_concurrency.lockutils [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Releasing lock "refresh_cache-443e235d-32b2-4af7-bdae-3f30e8196ee3" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 608.150018] env[62460]: DEBUG nova.compute.manager [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 608.150018] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 608.150018] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-73bc16d8-0d42-4338-9720-f74975ff18c7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.157954] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd9961f8-f7a1-4d25-9fba-0bcbb9980c83 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.185407] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 443e235d-32b2-4af7-bdae-3f30e8196ee3 could not be found. [ 608.186040] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 608.186388] env[62460]: INFO nova.compute.manager [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] Took 0.04 seconds to destroy the instance on the hypervisor. [ 608.186761] env[62460]: DEBUG oslo.service.loopingcall [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 608.189869] env[62460]: DEBUG nova.compute.manager [-] [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 608.191661] env[62460]: DEBUG nova.network.neutron [-] [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 608.235844] env[62460]: DEBUG nova.network.neutron [-] [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 608.428223] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-780bd604-76f2-479b-a5b1-cc7037ed3784 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.436748] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-930fe3fb-78fb-4121-8b1a-30cf27dab9e7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.469744] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b6521d2-ee6d-41db-b16e-b629a31a1a6b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.478460] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37d0d542-057c-4d64-ac37-8cebd43c32aa {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.494369] env[62460]: DEBUG nova.compute.provider_tree [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 608.739899] env[62460]: DEBUG nova.network.neutron [-] [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.818874] env[62460]: INFO nova.scheduler.client.report [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Deleted allocations for instance b51ddc3a-de97-4d51-adc9-e92b4fb0a150 [ 609.001066] env[62460]: DEBUG nova.scheduler.client.report [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 609.242384] env[62460]: INFO nova.compute.manager [-] [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] Took 1.05 seconds to deallocate network for instance. [ 609.248203] env[62460]: DEBUG nova.compute.claims [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 609.248203] env[62460]: DEBUG oslo_concurrency.lockutils [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 609.332585] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7d30f389-7fab-4aec-b4fa-03f1e5e9089b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "b51ddc3a-de97-4d51-adc9-e92b4fb0a150" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 62.010s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 609.510292] env[62460]: DEBUG oslo_concurrency.lockutils [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.999s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 609.510958] env[62460]: ERROR nova.compute.manager [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a9063309-85d1-4e90-8756-635659c53ff6, please check neutron logs for more information. [ 609.510958] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] Traceback (most recent call last): [ 609.510958] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 609.510958] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] self.driver.spawn(context, instance, image_meta, [ 609.510958] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 609.510958] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] self._vmops.spawn(context, instance, image_meta, injected_files, [ 609.510958] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 609.510958] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] vm_ref = self.build_virtual_machine(instance, [ 609.510958] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 609.510958] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] vif_infos = vmwarevif.get_vif_info(self._session, [ 609.510958] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 609.511256] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] for vif in network_info: [ 609.511256] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 609.511256] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] return self._sync_wrapper(fn, *args, **kwargs) [ 609.511256] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 609.511256] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] self.wait() [ 609.511256] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 609.511256] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] self[:] = self._gt.wait() [ 609.511256] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 609.511256] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] return self._exit_event.wait() [ 609.511256] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 609.511256] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] result = hub.switch() [ 609.511256] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 609.511256] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] return self.greenlet.switch() [ 609.511623] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 609.511623] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] result = function(*args, **kwargs) [ 609.511623] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 609.511623] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] return func(*args, **kwargs) [ 609.511623] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 609.511623] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] raise e [ 609.511623] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 609.511623] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] nwinfo = self.network_api.allocate_for_instance( [ 609.511623] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 609.511623] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] created_port_ids = self._update_ports_for_instance( [ 609.511623] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 609.511623] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] with excutils.save_and_reraise_exception(): [ 609.511623] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 609.511911] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] self.force_reraise() [ 609.511911] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 609.511911] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] raise self.value [ 609.511911] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 609.511911] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] updated_port = self._update_port( [ 609.511911] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 609.511911] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] _ensure_no_port_binding_failure(port) [ 609.511911] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 609.511911] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] raise exception.PortBindingFailed(port_id=port['id']) [ 609.511911] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] nova.exception.PortBindingFailed: Binding failed for port a9063309-85d1-4e90-8756-635659c53ff6, please check neutron logs for more information. [ 609.511911] env[62460]: ERROR nova.compute.manager [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] [ 609.512263] env[62460]: DEBUG nova.compute.utils [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] Binding failed for port a9063309-85d1-4e90-8756-635659c53ff6, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 609.512973] env[62460]: DEBUG oslo_concurrency.lockutils [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.976s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 609.521016] env[62460]: DEBUG nova.compute.manager [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] Build of instance 6ea9a22a-d6ca-4e14-b83e-97074d4cb531 was re-scheduled: Binding failed for port a9063309-85d1-4e90-8756-635659c53ff6, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 609.521016] env[62460]: DEBUG nova.compute.manager [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 609.521016] env[62460]: DEBUG oslo_concurrency.lockutils [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Acquiring lock "refresh_cache-6ea9a22a-d6ca-4e14-b83e-97074d4cb531" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 609.521016] env[62460]: DEBUG oslo_concurrency.lockutils [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Acquired lock "refresh_cache-6ea9a22a-d6ca-4e14-b83e-97074d4cb531" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 609.521361] env[62460]: DEBUG nova.network.neutron [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 609.840451] env[62460]: DEBUG nova.compute.manager [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] [instance: 94703126-bb31-40af-b945-2ef5ff37e094] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 610.046979] env[62460]: DEBUG nova.network.neutron [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 610.177353] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] Acquiring lock "2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 610.177479] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] Lock "2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 610.227490] env[62460]: DEBUG nova.network.neutron [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.370990] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 610.487681] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39b40d88-545f-4d37-93c7-907976a2609f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.501497] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0b6067e-cd92-4ab1-9736-b78e2f380dbd {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.533048] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b293e0c-ac2c-4c20-a464-b6dfa0acd1be {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.543226] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cf74f92-6a3a-467f-bca4-383e24fa349f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.558118] env[62460]: DEBUG nova.compute.provider_tree [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 610.733255] env[62460]: DEBUG oslo_concurrency.lockutils [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Releasing lock "refresh_cache-6ea9a22a-d6ca-4e14-b83e-97074d4cb531" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 610.733520] env[62460]: DEBUG nova.compute.manager [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 610.733710] env[62460]: DEBUG nova.compute.manager [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 610.733880] env[62460]: DEBUG nova.network.neutron [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 610.749467] env[62460]: DEBUG nova.network.neutron [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 610.934814] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] Acquiring lock "c37f1b55-7a06-4c93-9492-d2bd6d2bfe83" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 610.935093] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] Lock "c37f1b55-7a06-4c93-9492-d2bd6d2bfe83" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 611.061182] env[62460]: DEBUG nova.scheduler.client.report [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 611.177500] env[62460]: DEBUG oslo_concurrency.lockutils [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "ec2297bb-1d29-4ea9-90f7-f4f39716c103" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 611.177500] env[62460]: DEBUG oslo_concurrency.lockutils [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "ec2297bb-1d29-4ea9-90f7-f4f39716c103" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 611.256714] env[62460]: DEBUG nova.network.neutron [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.557819] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] Acquiring lock "0e00ca48-f9e1-4999-aad5-d7965f7ddc28" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 611.558152] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] Lock "0e00ca48-f9e1-4999-aad5-d7965f7ddc28" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 611.565365] env[62460]: DEBUG oslo_concurrency.lockutils [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.052s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 611.565984] env[62460]: ERROR nova.compute.manager [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port fbb7d4a3-775b-4aac-aa13-f88699a16b64, please check neutron logs for more information. [ 611.565984] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] Traceback (most recent call last): [ 611.565984] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 611.565984] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] self.driver.spawn(context, instance, image_meta, [ 611.565984] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 611.565984] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] self._vmops.spawn(context, instance, image_meta, injected_files, [ 611.565984] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 611.565984] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] vm_ref = self.build_virtual_machine(instance, [ 611.565984] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 611.565984] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] vif_infos = vmwarevif.get_vif_info(self._session, [ 611.565984] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 611.566284] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] for vif in network_info: [ 611.566284] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 611.566284] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] return self._sync_wrapper(fn, *args, **kwargs) [ 611.566284] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 611.566284] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] self.wait() [ 611.566284] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 611.566284] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] self[:] = self._gt.wait() [ 611.566284] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 611.566284] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] return self._exit_event.wait() [ 611.566284] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 611.566284] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] result = hub.switch() [ 611.566284] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 611.566284] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] return self.greenlet.switch() [ 611.566600] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 611.566600] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] result = function(*args, **kwargs) [ 611.566600] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 611.566600] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] return func(*args, **kwargs) [ 611.566600] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 611.566600] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] raise e [ 611.566600] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 611.566600] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] nwinfo = self.network_api.allocate_for_instance( [ 611.566600] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 611.566600] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] created_port_ids = self._update_ports_for_instance( [ 611.566600] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 611.566600] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] with excutils.save_and_reraise_exception(): [ 611.566600] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.566910] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] self.force_reraise() [ 611.566910] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.566910] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] raise self.value [ 611.566910] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 611.566910] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] updated_port = self._update_port( [ 611.566910] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.566910] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] _ensure_no_port_binding_failure(port) [ 611.566910] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.566910] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] raise exception.PortBindingFailed(port_id=port['id']) [ 611.566910] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] nova.exception.PortBindingFailed: Binding failed for port fbb7d4a3-775b-4aac-aa13-f88699a16b64, please check neutron logs for more information. [ 611.566910] env[62460]: ERROR nova.compute.manager [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] [ 611.567185] env[62460]: DEBUG nova.compute.utils [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] Binding failed for port fbb7d4a3-775b-4aac-aa13-f88699a16b64, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 611.567827] env[62460]: DEBUG oslo_concurrency.lockutils [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.490s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 611.570729] env[62460]: DEBUG nova.compute.manager [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] Build of instance 291beb16-db0e-4eb6-a224-2acfc14d6d77 was re-scheduled: Binding failed for port fbb7d4a3-775b-4aac-aa13-f88699a16b64, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 611.573436] env[62460]: DEBUG nova.compute.manager [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 611.573672] env[62460]: DEBUG oslo_concurrency.lockutils [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] Acquiring lock "refresh_cache-291beb16-db0e-4eb6-a224-2acfc14d6d77" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 611.573823] env[62460]: DEBUG oslo_concurrency.lockutils [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] Acquired lock "refresh_cache-291beb16-db0e-4eb6-a224-2acfc14d6d77" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 611.573996] env[62460]: DEBUG nova.network.neutron [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 611.759884] env[62460]: INFO nova.compute.manager [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 6ea9a22a-d6ca-4e14-b83e-97074d4cb531] Took 1.03 seconds to deallocate network for instance. [ 612.101888] env[62460]: DEBUG nova.network.neutron [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 612.225909] env[62460]: DEBUG nova.network.neutron [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.495287] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13c12638-a668-4342-b3a9-cce05f49d2f9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.503722] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e342a7a-3671-4d5a-ba0f-fefa3d29b582 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.535856] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae6f5f6f-322d-4dfd-bc4c-f44d5b930161 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.544171] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-907f279b-3511-4062-b457-d19357c87067 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.558437] env[62460]: DEBUG nova.compute.provider_tree [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 612.731447] env[62460]: DEBUG oslo_concurrency.lockutils [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] Releasing lock "refresh_cache-291beb16-db0e-4eb6-a224-2acfc14d6d77" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 612.731690] env[62460]: DEBUG nova.compute.manager [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 612.731875] env[62460]: DEBUG nova.compute.manager [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 612.732103] env[62460]: DEBUG nova.network.neutron [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 612.750672] env[62460]: DEBUG nova.network.neutron [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 612.791974] env[62460]: INFO nova.scheduler.client.report [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Deleted allocations for instance 6ea9a22a-d6ca-4e14-b83e-97074d4cb531 [ 613.062194] env[62460]: DEBUG nova.scheduler.client.report [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 613.253737] env[62460]: DEBUG nova.network.neutron [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.299703] env[62460]: DEBUG oslo_concurrency.lockutils [None req-bb876912-e979-4092-b614-b332b55b4693 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Lock "6ea9a22a-d6ca-4e14-b83e-97074d4cb531" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 65.624s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 613.567053] env[62460]: DEBUG oslo_concurrency.lockutils [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.999s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 613.567703] env[62460]: ERROR nova.compute.manager [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 14e5de25-c06c-48c1-a690-db4605dc0193, please check neutron logs for more information. [ 613.567703] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] Traceback (most recent call last): [ 613.567703] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 613.567703] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] self.driver.spawn(context, instance, image_meta, [ 613.567703] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 613.567703] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] self._vmops.spawn(context, instance, image_meta, injected_files, [ 613.567703] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 613.567703] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] vm_ref = self.build_virtual_machine(instance, [ 613.567703] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 613.567703] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] vif_infos = vmwarevif.get_vif_info(self._session, [ 613.567703] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 613.567986] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] for vif in network_info: [ 613.567986] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 613.567986] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] return self._sync_wrapper(fn, *args, **kwargs) [ 613.567986] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 613.567986] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] self.wait() [ 613.567986] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 613.567986] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] self[:] = self._gt.wait() [ 613.567986] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 613.567986] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] return self._exit_event.wait() [ 613.567986] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 613.567986] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] result = hub.switch() [ 613.567986] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 613.567986] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] return self.greenlet.switch() [ 613.568353] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 613.568353] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] result = function(*args, **kwargs) [ 613.568353] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 613.568353] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] return func(*args, **kwargs) [ 613.568353] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 613.568353] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] raise e [ 613.568353] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 613.568353] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] nwinfo = self.network_api.allocate_for_instance( [ 613.568353] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 613.568353] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] created_port_ids = self._update_ports_for_instance( [ 613.568353] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 613.568353] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] with excutils.save_and_reraise_exception(): [ 613.568353] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 613.568644] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] self.force_reraise() [ 613.568644] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 613.568644] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] raise self.value [ 613.568644] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 613.568644] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] updated_port = self._update_port( [ 613.568644] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 613.568644] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] _ensure_no_port_binding_failure(port) [ 613.568644] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 613.568644] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] raise exception.PortBindingFailed(port_id=port['id']) [ 613.568644] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] nova.exception.PortBindingFailed: Binding failed for port 14e5de25-c06c-48c1-a690-db4605dc0193, please check neutron logs for more information. [ 613.568644] env[62460]: ERROR nova.compute.manager [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] [ 613.568887] env[62460]: DEBUG nova.compute.utils [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] Binding failed for port 14e5de25-c06c-48c1-a690-db4605dc0193, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 613.569749] env[62460]: DEBUG oslo_concurrency.lockutils [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.830s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 613.573657] env[62460]: DEBUG nova.compute.manager [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] Build of instance 9b8ff88e-9e13-42b2-92b0-13af6873c916 was re-scheduled: Binding failed for port 14e5de25-c06c-48c1-a690-db4605dc0193, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 613.574139] env[62460]: DEBUG nova.compute.manager [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 613.574372] env[62460]: DEBUG oslo_concurrency.lockutils [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Acquiring lock "refresh_cache-9b8ff88e-9e13-42b2-92b0-13af6873c916" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 613.574523] env[62460]: DEBUG oslo_concurrency.lockutils [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Acquired lock "refresh_cache-9b8ff88e-9e13-42b2-92b0-13af6873c916" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 613.574685] env[62460]: DEBUG nova.network.neutron [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 613.756640] env[62460]: INFO nova.compute.manager [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] [instance: 291beb16-db0e-4eb6-a224-2acfc14d6d77] Took 1.02 seconds to deallocate network for instance. [ 613.802788] env[62460]: DEBUG nova.compute.manager [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 614.107386] env[62460]: DEBUG nova.network.neutron [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 614.205163] env[62460]: DEBUG nova.network.neutron [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.325507] env[62460]: DEBUG oslo_concurrency.lockutils [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 614.418015] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23a96c09-76b5-40a2-a286-af1f98ea2df5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.425849] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd096990-0a1d-4c05-a860-84991a353281 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.456754] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6597cae-d62e-4a37-9c50-329d682a3aae {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.463812] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62390909-faab-4c40-9062-16b57934f8ae {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.477077] env[62460]: DEBUG nova.compute.provider_tree [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 614.708650] env[62460]: DEBUG oslo_concurrency.lockutils [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Releasing lock "refresh_cache-9b8ff88e-9e13-42b2-92b0-13af6873c916" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 614.708939] env[62460]: DEBUG nova.compute.manager [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 614.709130] env[62460]: DEBUG nova.compute.manager [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 614.709306] env[62460]: DEBUG nova.network.neutron [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 614.733673] env[62460]: DEBUG nova.network.neutron [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 614.787020] env[62460]: INFO nova.scheduler.client.report [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] Deleted allocations for instance 291beb16-db0e-4eb6-a224-2acfc14d6d77 [ 614.979775] env[62460]: DEBUG nova.scheduler.client.report [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 615.236343] env[62460]: DEBUG nova.network.neutron [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 615.294367] env[62460]: DEBUG oslo_concurrency.lockutils [None req-8cd8c1f5-08d6-4f6b-b522-60e6425e268e tempest-InstanceActionsTestJSON-1030755183 tempest-InstanceActionsTestJSON-1030755183-project-member] Lock "291beb16-db0e-4eb6-a224-2acfc14d6d77" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 66.084s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 615.484455] env[62460]: DEBUG oslo_concurrency.lockutils [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.915s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 615.485143] env[62460]: ERROR nova.compute.manager [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 536f1463-299b-4c7b-8962-741db114114c, please check neutron logs for more information. [ 615.485143] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] Traceback (most recent call last): [ 615.485143] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 615.485143] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] self.driver.spawn(context, instance, image_meta, [ 615.485143] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 615.485143] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] self._vmops.spawn(context, instance, image_meta, injected_files, [ 615.485143] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 615.485143] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] vm_ref = self.build_virtual_machine(instance, [ 615.485143] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 615.485143] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] vif_infos = vmwarevif.get_vif_info(self._session, [ 615.485143] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 615.485415] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] for vif in network_info: [ 615.485415] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 615.485415] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] return self._sync_wrapper(fn, *args, **kwargs) [ 615.485415] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 615.485415] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] self.wait() [ 615.485415] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 615.485415] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] self[:] = self._gt.wait() [ 615.485415] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 615.485415] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] return self._exit_event.wait() [ 615.485415] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 615.485415] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] result = hub.switch() [ 615.485415] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 615.485415] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] return self.greenlet.switch() [ 615.485713] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 615.485713] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] result = function(*args, **kwargs) [ 615.485713] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 615.485713] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] return func(*args, **kwargs) [ 615.485713] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 615.485713] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] raise e [ 615.485713] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 615.485713] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] nwinfo = self.network_api.allocate_for_instance( [ 615.485713] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 615.485713] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] created_port_ids = self._update_ports_for_instance( [ 615.485713] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 615.485713] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] with excutils.save_and_reraise_exception(): [ 615.485713] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 615.486537] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] self.force_reraise() [ 615.486537] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 615.486537] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] raise self.value [ 615.486537] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 615.486537] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] updated_port = self._update_port( [ 615.486537] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 615.486537] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] _ensure_no_port_binding_failure(port) [ 615.486537] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 615.486537] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] raise exception.PortBindingFailed(port_id=port['id']) [ 615.486537] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] nova.exception.PortBindingFailed: Binding failed for port 536f1463-299b-4c7b-8962-741db114114c, please check neutron logs for more information. [ 615.486537] env[62460]: ERROR nova.compute.manager [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] [ 615.486983] env[62460]: DEBUG nova.compute.utils [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] Binding failed for port 536f1463-299b-4c7b-8962-741db114114c, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 615.487113] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.421s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 615.489051] env[62460]: INFO nova.compute.claims [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] [instance: 3b71c366-cbd9-4b98-aa0d-c55b56e69231] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 615.491222] env[62460]: DEBUG nova.compute.manager [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] Build of instance a8d7055e-e7a3-470a-ae96-f89753afa9df was re-scheduled: Binding failed for port 536f1463-299b-4c7b-8962-741db114114c, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 615.491655] env[62460]: DEBUG nova.compute.manager [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 615.491880] env[62460]: DEBUG oslo_concurrency.lockutils [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] Acquiring lock "refresh_cache-a8d7055e-e7a3-470a-ae96-f89753afa9df" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 615.492062] env[62460]: DEBUG oslo_concurrency.lockutils [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] Acquired lock "refresh_cache-a8d7055e-e7a3-470a-ae96-f89753afa9df" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 615.492203] env[62460]: DEBUG nova.network.neutron [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 615.738880] env[62460]: INFO nova.compute.manager [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] [instance: 9b8ff88e-9e13-42b2-92b0-13af6873c916] Took 1.03 seconds to deallocate network for instance. [ 615.797315] env[62460]: DEBUG nova.compute.manager [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 616.011274] env[62460]: DEBUG nova.network.neutron [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 616.101092] env[62460]: DEBUG nova.network.neutron [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.324905] env[62460]: DEBUG oslo_concurrency.lockutils [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 616.605506] env[62460]: DEBUG oslo_concurrency.lockutils [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] Releasing lock "refresh_cache-a8d7055e-e7a3-470a-ae96-f89753afa9df" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 616.605738] env[62460]: DEBUG nova.compute.manager [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 616.605916] env[62460]: DEBUG nova.compute.manager [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 616.606100] env[62460]: DEBUG nova.network.neutron [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 616.623590] env[62460]: DEBUG nova.network.neutron [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 616.771026] env[62460]: INFO nova.scheduler.client.report [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Deleted allocations for instance 9b8ff88e-9e13-42b2-92b0-13af6873c916 [ 616.868351] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b8ebf4f-8985-409b-9c3c-8e67c17a6c5d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.877196] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a9a64e8-56d1-4199-a166-e918b6915576 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.910070] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f205e14-8767-4ba4-bf2d-11685829c3ef {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.918600] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-429f2170-69c6-4e08-a8d3-3dc2214444fd {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.939028] env[62460]: DEBUG nova.compute.provider_tree [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 617.133863] env[62460]: DEBUG nova.network.neutron [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.281722] env[62460]: DEBUG oslo_concurrency.lockutils [None req-71cda14e-1f81-44e8-a401-5d72a8af2046 tempest-ServersAdminTestJSON-1906789592 tempest-ServersAdminTestJSON-1906789592-project-member] Lock "9b8ff88e-9e13-42b2-92b0-13af6873c916" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 67.557s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 617.442255] env[62460]: DEBUG nova.scheduler.client.report [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 617.637316] env[62460]: INFO nova.compute.manager [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] [instance: a8d7055e-e7a3-470a-ae96-f89753afa9df] Took 1.03 seconds to deallocate network for instance. [ 617.784385] env[62460]: DEBUG nova.compute.manager [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 617.948127] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.460s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 617.948127] env[62460]: DEBUG nova.compute.manager [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] [instance: 3b71c366-cbd9-4b98-aa0d-c55b56e69231] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 617.950204] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 14.953s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 617.950388] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 617.950546] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62460) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 617.950827] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.905s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 617.956124] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebc1f4ce-568a-4846-b9b4-bbf0955ee358 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.967017] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7411a5b5-b312-4e18-bc28-ecb6c7ca7b2e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.985247] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26d939f1-1e7f-437b-8d1f-d55ebc4f12db {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.989576] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb78fc28-0aa5-4f55-8983-869c8a34c27a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.020214] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181315MB free_disk=135GB free_vcpus=48 pci_devices=None {{(pid=62460) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 618.020393] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 618.307013] env[62460]: DEBUG oslo_concurrency.lockutils [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 618.460225] env[62460]: DEBUG nova.compute.utils [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 618.461769] env[62460]: DEBUG nova.compute.manager [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] [instance: 3b71c366-cbd9-4b98-aa0d-c55b56e69231] Not allocating networking since 'none' was specified. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 618.668851] env[62460]: INFO nova.scheduler.client.report [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] Deleted allocations for instance a8d7055e-e7a3-470a-ae96-f89753afa9df [ 618.819504] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d164dda-3361-4c8c-acae-2a88d4f3fb3f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.827493] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c1f9e37-60ed-4a22-b582-83b6eea9c3cc {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.858061] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cef6a8e-a2e0-4202-95e2-2cf1469e5120 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.870553] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f48f7555-e793-486c-ad0d-3c241423dc76 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.883883] env[62460]: DEBUG nova.compute.provider_tree [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 618.962983] env[62460]: DEBUG nova.compute.manager [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] [instance: 3b71c366-cbd9-4b98-aa0d-c55b56e69231] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 619.175224] env[62460]: DEBUG oslo_concurrency.lockutils [None req-000a6e05-c29d-431d-b3cf-e1b766bbd3c5 tempest-ServerRescueTestJSON-381716336 tempest-ServerRescueTestJSON-381716336-project-member] Lock "a8d7055e-e7a3-470a-ae96-f89753afa9df" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 66.741s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 619.388054] env[62460]: DEBUG nova.scheduler.client.report [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 619.678291] env[62460]: DEBUG nova.compute.manager [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 619.894990] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.942s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 619.894990] env[62460]: ERROR nova.compute.manager [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] [instance: a2747204-95c5-4200-8742-9ec39b4368c8] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 660a85f6-adeb-4e4c-be35-ae655ea19cfd, please check neutron logs for more information. [ 619.894990] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] Traceback (most recent call last): [ 619.894990] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 619.894990] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] self.driver.spawn(context, instance, image_meta, [ 619.894990] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 619.894990] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 619.894990] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 619.894990] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] vm_ref = self.build_virtual_machine(instance, [ 619.895475] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 619.895475] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] vif_infos = vmwarevif.get_vif_info(self._session, [ 619.895475] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 619.895475] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] for vif in network_info: [ 619.895475] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 619.895475] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] return self._sync_wrapper(fn, *args, **kwargs) [ 619.895475] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 619.895475] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] self.wait() [ 619.895475] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 619.895475] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] self[:] = self._gt.wait() [ 619.895475] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 619.895475] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] return self._exit_event.wait() [ 619.895475] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 619.895777] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] current.throw(*self._exc) [ 619.895777] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 619.895777] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] result = function(*args, **kwargs) [ 619.895777] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 619.895777] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] return func(*args, **kwargs) [ 619.895777] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 619.895777] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] raise e [ 619.895777] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 619.895777] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] nwinfo = self.network_api.allocate_for_instance( [ 619.895777] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 619.895777] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] created_port_ids = self._update_ports_for_instance( [ 619.895777] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 619.895777] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] with excutils.save_and_reraise_exception(): [ 619.896128] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 619.896128] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] self.force_reraise() [ 619.896128] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 619.896128] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] raise self.value [ 619.896128] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 619.896128] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] updated_port = self._update_port( [ 619.896128] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 619.896128] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] _ensure_no_port_binding_failure(port) [ 619.896128] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 619.896128] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] raise exception.PortBindingFailed(port_id=port['id']) [ 619.896128] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] nova.exception.PortBindingFailed: Binding failed for port 660a85f6-adeb-4e4c-be35-ae655ea19cfd, please check neutron logs for more information. [ 619.896128] env[62460]: ERROR nova.compute.manager [instance: a2747204-95c5-4200-8742-9ec39b4368c8] [ 619.896405] env[62460]: DEBUG nova.compute.utils [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] [instance: a2747204-95c5-4200-8742-9ec39b4368c8] Binding failed for port 660a85f6-adeb-4e4c-be35-ae655ea19cfd, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 619.897030] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.818s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 619.897984] env[62460]: INFO nova.compute.claims [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 619.905352] env[62460]: DEBUG nova.compute.manager [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] [instance: a2747204-95c5-4200-8742-9ec39b4368c8] Build of instance a2747204-95c5-4200-8742-9ec39b4368c8 was re-scheduled: Binding failed for port 660a85f6-adeb-4e4c-be35-ae655ea19cfd, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 619.905352] env[62460]: DEBUG nova.compute.manager [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] [instance: a2747204-95c5-4200-8742-9ec39b4368c8] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 619.905352] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] Acquiring lock "refresh_cache-a2747204-95c5-4200-8742-9ec39b4368c8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 619.905352] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] Acquired lock "refresh_cache-a2747204-95c5-4200-8742-9ec39b4368c8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 619.905580] env[62460]: DEBUG nova.network.neutron [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] [instance: a2747204-95c5-4200-8742-9ec39b4368c8] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 619.972806] env[62460]: DEBUG nova.compute.manager [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] [instance: 3b71c366-cbd9-4b98-aa0d-c55b56e69231] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 620.004916] env[62460]: DEBUG nova.virt.hardware [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 620.005184] env[62460]: DEBUG nova.virt.hardware [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 620.005573] env[62460]: DEBUG nova.virt.hardware [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 620.005869] env[62460]: DEBUG nova.virt.hardware [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 620.006086] env[62460]: DEBUG nova.virt.hardware [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 620.006304] env[62460]: DEBUG nova.virt.hardware [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 620.006585] env[62460]: DEBUG nova.virt.hardware [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 620.007568] env[62460]: DEBUG nova.virt.hardware [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 620.007912] env[62460]: DEBUG nova.virt.hardware [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 620.008399] env[62460]: DEBUG nova.virt.hardware [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 620.008691] env[62460]: DEBUG nova.virt.hardware [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 620.012652] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8243e1f9-1761-444a-9370-65138bab50a2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.022809] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2d37249-3e8d-4ad6-930e-f6d538d9b2e0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.040188] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] [instance: 3b71c366-cbd9-4b98-aa0d-c55b56e69231] Instance VIF info [] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 620.046624] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Creating folder: Project (2275af3492d743f1922bd2c55cf9a07e). Parent ref: group-v281134. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 620.047186] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4fa5adee-1e83-4a2f-88a3-77868f130c06 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.058366] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Created folder: Project (2275af3492d743f1922bd2c55cf9a07e) in parent group-v281134. [ 620.058738] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Creating folder: Instances. Parent ref: group-v281147. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 620.059129] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5f028a5d-0cdd-4515-8893-b3c09414c59b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.069416] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Created folder: Instances in parent group-v281147. [ 620.069501] env[62460]: DEBUG oslo.service.loopingcall [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 620.069670] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3b71c366-cbd9-4b98-aa0d-c55b56e69231] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 620.069858] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c1da020d-42c6-4aef-96cf-c486ce7be2be {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.088612] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 620.088612] env[62460]: value = "task-1313439" [ 620.088612] env[62460]: _type = "Task" [ 620.088612] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.097045] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313439, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.207830] env[62460]: DEBUG oslo_concurrency.lockutils [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 620.427524] env[62460]: DEBUG nova.network.neutron [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] [instance: a2747204-95c5-4200-8742-9ec39b4368c8] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 620.560944] env[62460]: DEBUG nova.network.neutron [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] [instance: a2747204-95c5-4200-8742-9ec39b4368c8] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.598472] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313439, 'name': CreateVM_Task, 'duration_secs': 0.260656} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 620.601601] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3b71c366-cbd9-4b98-aa0d-c55b56e69231] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 620.601601] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 620.601601] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 620.601601] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 620.601601] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b4d728c-8b10-4dbc-b470-9083942b7b2f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.606904] env[62460]: DEBUG oslo_vmware.api [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Waiting for the task: (returnval){ [ 620.606904] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]528feddd-cdad-497c-a801-3736e12a2e4a" [ 620.606904] env[62460]: _type = "Task" [ 620.606904] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.611753] env[62460]: DEBUG oslo_vmware.api [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]528feddd-cdad-497c-a801-3736e12a2e4a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.064294] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] Releasing lock "refresh_cache-a2747204-95c5-4200-8742-9ec39b4368c8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 621.065192] env[62460]: DEBUG nova.compute.manager [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 621.065192] env[62460]: DEBUG nova.compute.manager [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] [instance: a2747204-95c5-4200-8742-9ec39b4368c8] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 621.065192] env[62460]: DEBUG nova.network.neutron [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] [instance: a2747204-95c5-4200-8742-9ec39b4368c8] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 621.085901] env[62460]: DEBUG nova.network.neutron [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] [instance: a2747204-95c5-4200-8742-9ec39b4368c8] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 621.118992] env[62460]: DEBUG oslo_vmware.api [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]528feddd-cdad-497c-a801-3736e12a2e4a, 'name': SearchDatastore_Task, 'duration_secs': 0.0111} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.119330] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 621.119564] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] [instance: 3b71c366-cbd9-4b98-aa0d-c55b56e69231] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 621.119793] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 621.120268] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 621.120268] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 621.120376] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4058b2cd-2568-4079-8fbb-98b9c7b36399 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.134654] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 621.134850] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 621.137387] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7677d9cf-dcb8-420c-84e0-995d9fd7d702 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.140894] env[62460]: DEBUG oslo_vmware.api [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Waiting for the task: (returnval){ [ 621.140894] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]528a8e0e-a167-92e4-df27-99fdebdbadeb" [ 621.140894] env[62460]: _type = "Task" [ 621.140894] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 621.151371] env[62460]: DEBUG oslo_vmware.api [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]528a8e0e-a167-92e4-df27-99fdebdbadeb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.330598] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11430e5f-3475-4b5e-9346-961128a3148a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.342230] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dff0e5c8-dc5e-4ab0-bdf1-b628846088ab {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.374105] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89d82f2f-35cb-47c3-bebb-5586c2c10e74 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.381239] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-658a6afa-0606-43a4-9e80-76dc9dd63770 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.394814] env[62460]: DEBUG nova.compute.provider_tree [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 621.589628] env[62460]: DEBUG nova.network.neutron [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] [instance: a2747204-95c5-4200-8742-9ec39b4368c8] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.651805] env[62460]: DEBUG oslo_vmware.api [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]528a8e0e-a167-92e4-df27-99fdebdbadeb, 'name': SearchDatastore_Task, 'duration_secs': 0.009334} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.652408] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-431b73a0-b3ea-439e-8e9d-7eacedab4d88 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.657345] env[62460]: DEBUG oslo_vmware.api [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Waiting for the task: (returnval){ [ 621.657345] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]5247598f-c6d2-3dc4-b869-c6de8353e3e8" [ 621.657345] env[62460]: _type = "Task" [ 621.657345] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 621.665129] env[62460]: DEBUG oslo_vmware.api [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5247598f-c6d2-3dc4-b869-c6de8353e3e8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.897951] env[62460]: DEBUG nova.scheduler.client.report [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 622.092280] env[62460]: INFO nova.compute.manager [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] [instance: a2747204-95c5-4200-8742-9ec39b4368c8] Took 1.03 seconds to deallocate network for instance. [ 622.168534] env[62460]: DEBUG oslo_vmware.api [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5247598f-c6d2-3dc4-b869-c6de8353e3e8, 'name': SearchDatastore_Task, 'duration_secs': 0.008918} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 622.168851] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 622.169150] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 3b71c366-cbd9-4b98-aa0d-c55b56e69231/3b71c366-cbd9-4b98-aa0d-c55b56e69231.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 622.169493] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e4203bbb-5856-4d24-b276-830637ebce28 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.176306] env[62460]: DEBUG oslo_vmware.api [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Waiting for the task: (returnval){ [ 622.176306] env[62460]: value = "task-1313440" [ 622.176306] env[62460]: _type = "Task" [ 622.176306] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 622.185679] env[62460]: DEBUG oslo_vmware.api [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Task: {'id': task-1313440, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.405874] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.509s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 622.407705] env[62460]: DEBUG nova.compute.manager [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 622.409619] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.289s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 622.411198] env[62460]: INFO nova.compute.claims [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 622.687910] env[62460]: DEBUG oslo_vmware.api [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Task: {'id': task-1313440, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.915938] env[62460]: DEBUG nova.compute.utils [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 622.919812] env[62460]: DEBUG nova.compute.manager [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 622.920063] env[62460]: DEBUG nova.network.neutron [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 623.131272] env[62460]: INFO nova.scheduler.client.report [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] Deleted allocations for instance a2747204-95c5-4200-8742-9ec39b4368c8 [ 623.157722] env[62460]: DEBUG nova.policy [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '632d7ca8cb4243f990d741a3f70f422a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '10ac54db1ef54e249a077bbb3dde4242', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 623.188769] env[62460]: DEBUG oslo_vmware.api [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Task: {'id': task-1313440, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.518416} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 623.189501] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 3b71c366-cbd9-4b98-aa0d-c55b56e69231/3b71c366-cbd9-4b98-aa0d-c55b56e69231.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 623.189766] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] [instance: 3b71c366-cbd9-4b98-aa0d-c55b56e69231] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 623.190051] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5f20b1c0-a96a-483b-bbb2-4260413fb5c8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.196986] env[62460]: DEBUG oslo_vmware.api [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Waiting for the task: (returnval){ [ 623.196986] env[62460]: value = "task-1313441" [ 623.196986] env[62460]: _type = "Task" [ 623.196986] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 623.204952] env[62460]: DEBUG oslo_vmware.api [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Task: {'id': task-1313441, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.420970] env[62460]: DEBUG nova.compute.manager [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 623.639815] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0e07be6c-4da7-4715-8106-147ebb86c2d8 tempest-ServerActionsV293TestJSON-854937809 tempest-ServerActionsV293TestJSON-854937809-project-member] Lock "a2747204-95c5-4200-8742-9ec39b4368c8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 69.106s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 623.649459] env[62460]: DEBUG nova.network.neutron [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] Successfully created port: 3373c77f-c40e-4658-9249-466f197b9940 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 623.711648] env[62460]: DEBUG oslo_vmware.api [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Task: {'id': task-1313441, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066412} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 623.714634] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] [instance: 3b71c366-cbd9-4b98-aa0d-c55b56e69231] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 623.716146] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4cda353-f05d-4f06-95ba-e9cba8864243 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.739604] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] [instance: 3b71c366-cbd9-4b98-aa0d-c55b56e69231] Reconfiguring VM instance instance-00000017 to attach disk [datastore1] 3b71c366-cbd9-4b98-aa0d-c55b56e69231/3b71c366-cbd9-4b98-aa0d-c55b56e69231.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 623.742856] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cd8e9a7a-1ec2-4123-b198-973d6cd55466 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.763589] env[62460]: DEBUG oslo_vmware.api [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Waiting for the task: (returnval){ [ 623.763589] env[62460]: value = "task-1313443" [ 623.763589] env[62460]: _type = "Task" [ 623.763589] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 623.774581] env[62460]: DEBUG oslo_vmware.api [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Task: {'id': task-1313443, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.829097] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b9e7b08-a8e3-4a1c-b385-b3d58c982d4e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.837794] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8ad2946-0c02-408e-a54b-b6d99859f8c4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.867446] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99c19da7-fe12-409a-a7f9-9c44f04f783c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.875550] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84426536-6b50-4153-8968-5bc56ed9f9d2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.889991] env[62460]: DEBUG nova.compute.provider_tree [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 624.145310] env[62460]: DEBUG nova.compute.manager [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 624.275596] env[62460]: DEBUG oslo_vmware.api [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Task: {'id': task-1313443, 'name': ReconfigVM_Task, 'duration_secs': 0.306073} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.276193] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] [instance: 3b71c366-cbd9-4b98-aa0d-c55b56e69231] Reconfigured VM instance instance-00000017 to attach disk [datastore1] 3b71c366-cbd9-4b98-aa0d-c55b56e69231/3b71c366-cbd9-4b98-aa0d-c55b56e69231.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 624.276628] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-210295cd-c7a5-48c7-9736-aa909924adb5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.284776] env[62460]: DEBUG oslo_vmware.api [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Waiting for the task: (returnval){ [ 624.284776] env[62460]: value = "task-1313444" [ 624.284776] env[62460]: _type = "Task" [ 624.284776] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.295269] env[62460]: DEBUG oslo_vmware.api [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Task: {'id': task-1313444, 'name': Rename_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.393516] env[62460]: DEBUG nova.scheduler.client.report [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 624.435653] env[62460]: DEBUG nova.compute.manager [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 624.470243] env[62460]: DEBUG nova.virt.hardware [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 624.470907] env[62460]: DEBUG nova.virt.hardware [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 624.471092] env[62460]: DEBUG nova.virt.hardware [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 624.471331] env[62460]: DEBUG nova.virt.hardware [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 624.471679] env[62460]: DEBUG nova.virt.hardware [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 624.471889] env[62460]: DEBUG nova.virt.hardware [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 624.472130] env[62460]: DEBUG nova.virt.hardware [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 624.472300] env[62460]: DEBUG nova.virt.hardware [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 624.472469] env[62460]: DEBUG nova.virt.hardware [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 624.472631] env[62460]: DEBUG nova.virt.hardware [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 624.473195] env[62460]: DEBUG nova.virt.hardware [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 624.474336] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9d46728-4c46-4df4-b054-594ce5263fa3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.487042] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8866a3d4-1283-41d9-bd35-6e883a5f0ae8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.673931] env[62460]: DEBUG oslo_concurrency.lockutils [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.801084] env[62460]: DEBUG oslo_vmware.api [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Task: {'id': task-1313444, 'name': Rename_Task, 'duration_secs': 0.128539} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.801084] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] [instance: 3b71c366-cbd9-4b98-aa0d-c55b56e69231] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 624.801084] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ad34cbaf-b074-4cc5-86e2-567490612a97 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.808829] env[62460]: DEBUG oslo_vmware.api [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Waiting for the task: (returnval){ [ 624.808829] env[62460]: value = "task-1313445" [ 624.808829] env[62460]: _type = "Task" [ 624.808829] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.814274] env[62460]: DEBUG oslo_vmware.api [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Task: {'id': task-1313445, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.901710] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.490s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 624.901710] env[62460]: DEBUG nova.compute.manager [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 624.904451] env[62460]: DEBUG oslo_concurrency.lockutils [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.658s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 625.224577] env[62460]: DEBUG nova.compute.manager [req-d95066c8-86ab-432c-ae3b-c3c495880107 req-8af269b9-3159-4431-83d4-abce72297564 service nova] [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] Received event network-changed-3373c77f-c40e-4658-9249-466f197b9940 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 625.224577] env[62460]: DEBUG nova.compute.manager [req-d95066c8-86ab-432c-ae3b-c3c495880107 req-8af269b9-3159-4431-83d4-abce72297564 service nova] [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] Refreshing instance network info cache due to event network-changed-3373c77f-c40e-4658-9249-466f197b9940. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 625.224577] env[62460]: DEBUG oslo_concurrency.lockutils [req-d95066c8-86ab-432c-ae3b-c3c495880107 req-8af269b9-3159-4431-83d4-abce72297564 service nova] Acquiring lock "refresh_cache-9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 625.224577] env[62460]: DEBUG oslo_concurrency.lockutils [req-d95066c8-86ab-432c-ae3b-c3c495880107 req-8af269b9-3159-4431-83d4-abce72297564 service nova] Acquired lock "refresh_cache-9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.225517] env[62460]: DEBUG nova.network.neutron [req-d95066c8-86ab-432c-ae3b-c3c495880107 req-8af269b9-3159-4431-83d4-abce72297564 service nova] [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] Refreshing network info cache for port 3373c77f-c40e-4658-9249-466f197b9940 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 625.318617] env[62460]: DEBUG oslo_vmware.api [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Task: {'id': task-1313445, 'name': PowerOnVM_Task, 'duration_secs': 0.420939} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.323147] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] [instance: 3b71c366-cbd9-4b98-aa0d-c55b56e69231] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 625.323147] env[62460]: INFO nova.compute.manager [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] [instance: 3b71c366-cbd9-4b98-aa0d-c55b56e69231] Took 5.35 seconds to spawn the instance on the hypervisor. [ 625.323147] env[62460]: DEBUG nova.compute.manager [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] [instance: 3b71c366-cbd9-4b98-aa0d-c55b56e69231] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 625.323147] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf3cc40a-ec0c-4bd2-ac55-79334fd3db25 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.410283] env[62460]: DEBUG nova.compute.utils [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 625.411709] env[62460]: DEBUG nova.compute.manager [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 625.411884] env[62460]: DEBUG nova.network.neutron [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 625.438356] env[62460]: ERROR nova.compute.manager [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3373c77f-c40e-4658-9249-466f197b9940, please check neutron logs for more information. [ 625.438356] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 625.438356] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 625.438356] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 625.438356] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 625.438356] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 625.438356] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 625.438356] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 625.438356] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 625.438356] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 625.438356] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 625.438356] env[62460]: ERROR nova.compute.manager raise self.value [ 625.438356] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 625.438356] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 625.438356] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 625.438356] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 625.439189] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 625.439189] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 625.439189] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3373c77f-c40e-4658-9249-466f197b9940, please check neutron logs for more information. [ 625.439189] env[62460]: ERROR nova.compute.manager [ 625.439189] env[62460]: Traceback (most recent call last): [ 625.439189] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 625.439189] env[62460]: listener.cb(fileno) [ 625.439189] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 625.439189] env[62460]: result = function(*args, **kwargs) [ 625.439189] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 625.439189] env[62460]: return func(*args, **kwargs) [ 625.439189] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 625.439189] env[62460]: raise e [ 625.439189] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 625.439189] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 625.439189] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 625.439189] env[62460]: created_port_ids = self._update_ports_for_instance( [ 625.439189] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 625.439189] env[62460]: with excutils.save_and_reraise_exception(): [ 625.439189] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 625.439189] env[62460]: self.force_reraise() [ 625.439189] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 625.439189] env[62460]: raise self.value [ 625.439189] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 625.439189] env[62460]: updated_port = self._update_port( [ 625.439189] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 625.439189] env[62460]: _ensure_no_port_binding_failure(port) [ 625.439189] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 625.439189] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 625.440183] env[62460]: nova.exception.PortBindingFailed: Binding failed for port 3373c77f-c40e-4658-9249-466f197b9940, please check neutron logs for more information. [ 625.440183] env[62460]: Removing descriptor: 18 [ 625.440183] env[62460]: ERROR nova.compute.manager [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3373c77f-c40e-4658-9249-466f197b9940, please check neutron logs for more information. [ 625.440183] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] Traceback (most recent call last): [ 625.440183] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 625.440183] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] yield resources [ 625.440183] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 625.440183] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] self.driver.spawn(context, instance, image_meta, [ 625.440183] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 625.440183] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 625.440183] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 625.440183] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] vm_ref = self.build_virtual_machine(instance, [ 625.440492] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 625.440492] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] vif_infos = vmwarevif.get_vif_info(self._session, [ 625.440492] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 625.440492] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] for vif in network_info: [ 625.440492] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 625.440492] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] return self._sync_wrapper(fn, *args, **kwargs) [ 625.440492] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 625.440492] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] self.wait() [ 625.440492] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 625.440492] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] self[:] = self._gt.wait() [ 625.440492] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 625.440492] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] return self._exit_event.wait() [ 625.440492] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 625.441279] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] result = hub.switch() [ 625.441279] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 625.441279] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] return self.greenlet.switch() [ 625.441279] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 625.441279] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] result = function(*args, **kwargs) [ 625.441279] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 625.441279] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] return func(*args, **kwargs) [ 625.441279] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 625.441279] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] raise e [ 625.441279] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 625.441279] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] nwinfo = self.network_api.allocate_for_instance( [ 625.441279] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 625.441279] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] created_port_ids = self._update_ports_for_instance( [ 625.441606] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 625.441606] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] with excutils.save_and_reraise_exception(): [ 625.441606] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 625.441606] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] self.force_reraise() [ 625.441606] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 625.441606] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] raise self.value [ 625.441606] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 625.441606] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] updated_port = self._update_port( [ 625.441606] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 625.441606] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] _ensure_no_port_binding_failure(port) [ 625.441606] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 625.441606] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] raise exception.PortBindingFailed(port_id=port['id']) [ 625.442496] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] nova.exception.PortBindingFailed: Binding failed for port 3373c77f-c40e-4658-9249-466f197b9940, please check neutron logs for more information. [ 625.442496] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] [ 625.442496] env[62460]: INFO nova.compute.manager [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] Terminating instance [ 625.442496] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "refresh_cache-9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 625.484197] env[62460]: DEBUG nova.policy [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4d95c2a24f0841de8990410e24f325e9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '11491a5610734d96bac82035c46bd5d1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 625.747168] env[62460]: DEBUG nova.network.neutron [req-d95066c8-86ab-432c-ae3b-c3c495880107 req-8af269b9-3159-4431-83d4-abce72297564 service nova] [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 625.801017] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6491c9e2-7315-4908-a466-de1d0a7ca9be {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.807368] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b2238a2-df23-46d0-bee0-a84cb130b910 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.843972] env[62460]: DEBUG nova.network.neutron [req-d95066c8-86ab-432c-ae3b-c3c495880107 req-8af269b9-3159-4431-83d4-abce72297564 service nova] [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.848880] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-849255d8-7c9e-4c9a-bc09-18496572daf8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.860122] env[62460]: INFO nova.compute.manager [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] [instance: 3b71c366-cbd9-4b98-aa0d-c55b56e69231] Took 26.81 seconds to build instance. [ 625.864447] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e3ace7c-ba74-47e1-982b-d6a900be0316 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.883856] env[62460]: DEBUG nova.compute.provider_tree [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 625.915463] env[62460]: DEBUG nova.compute.manager [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 626.109654] env[62460]: DEBUG nova.network.neutron [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] Successfully created port: 3f617afc-843a-4ab7-88fe-7a981af30004 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 626.355778] env[62460]: DEBUG oslo_concurrency.lockutils [req-d95066c8-86ab-432c-ae3b-c3c495880107 req-8af269b9-3159-4431-83d4-abce72297564 service nova] Releasing lock "refresh_cache-9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 626.356242] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquired lock "refresh_cache-9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 626.356600] env[62460]: DEBUG nova.network.neutron [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 626.374471] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1bb344de-486c-4102-bb05-acff2c9191ca tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Lock "3b71c366-cbd9-4b98-aa0d-c55b56e69231" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 62.511s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 626.389121] env[62460]: DEBUG nova.scheduler.client.report [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 626.758632] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] Acquiring lock "ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 626.758894] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] Lock "ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 626.879330] env[62460]: DEBUG nova.network.neutron [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 626.882342] env[62460]: DEBUG nova.compute.manager [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] [instance: 61426715-7a38-475d-895a-0eb6d6040c66] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 626.895544] env[62460]: DEBUG oslo_concurrency.lockutils [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.991s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 626.897484] env[62460]: ERROR nova.compute.manager [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 63267a51-1e3d-4d36-ae79-2d9931707395, please check neutron logs for more information. [ 626.897484] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] Traceback (most recent call last): [ 626.897484] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 626.897484] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] self.driver.spawn(context, instance, image_meta, [ 626.897484] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 626.897484] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 626.897484] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 626.897484] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] vm_ref = self.build_virtual_machine(instance, [ 626.897484] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 626.897484] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] vif_infos = vmwarevif.get_vif_info(self._session, [ 626.897484] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 626.897853] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] for vif in network_info: [ 626.897853] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 626.897853] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] return self._sync_wrapper(fn, *args, **kwargs) [ 626.897853] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 626.897853] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] self.wait() [ 626.897853] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 626.897853] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] self[:] = self._gt.wait() [ 626.897853] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 626.897853] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] return self._exit_event.wait() [ 626.897853] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 626.897853] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] result = hub.switch() [ 626.897853] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 626.897853] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] return self.greenlet.switch() [ 626.898330] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 626.898330] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] result = function(*args, **kwargs) [ 626.898330] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 626.898330] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] return func(*args, **kwargs) [ 626.898330] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 626.898330] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] raise e [ 626.898330] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 626.898330] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] nwinfo = self.network_api.allocate_for_instance( [ 626.898330] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 626.898330] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] created_port_ids = self._update_ports_for_instance( [ 626.898330] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 626.898330] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] with excutils.save_and_reraise_exception(): [ 626.898330] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 626.898683] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] self.force_reraise() [ 626.898683] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 626.898683] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] raise self.value [ 626.898683] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 626.898683] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] updated_port = self._update_port( [ 626.898683] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 626.898683] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] _ensure_no_port_binding_failure(port) [ 626.898683] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 626.898683] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] raise exception.PortBindingFailed(port_id=port['id']) [ 626.898683] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] nova.exception.PortBindingFailed: Binding failed for port 63267a51-1e3d-4d36-ae79-2d9931707395, please check neutron logs for more information. [ 626.898683] env[62460]: ERROR nova.compute.manager [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] [ 626.898987] env[62460]: DEBUG nova.compute.utils [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] Binding failed for port 63267a51-1e3d-4d36-ae79-2d9931707395, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 626.901178] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.530s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 626.902170] env[62460]: INFO nova.compute.claims [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] [instance: 94703126-bb31-40af-b945-2ef5ff37e094] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 626.905862] env[62460]: DEBUG nova.compute.manager [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] Build of instance 443e235d-32b2-4af7-bdae-3f30e8196ee3 was re-scheduled: Binding failed for port 63267a51-1e3d-4d36-ae79-2d9931707395, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 626.905862] env[62460]: DEBUG nova.compute.manager [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 626.905862] env[62460]: DEBUG oslo_concurrency.lockutils [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Acquiring lock "refresh_cache-443e235d-32b2-4af7-bdae-3f30e8196ee3" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 626.906012] env[62460]: DEBUG oslo_concurrency.lockutils [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Acquired lock "refresh_cache-443e235d-32b2-4af7-bdae-3f30e8196ee3" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 626.906123] env[62460]: DEBUG nova.network.neutron [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 626.924989] env[62460]: DEBUG nova.compute.manager [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 626.961563] env[62460]: DEBUG nova.virt.hardware [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 626.961939] env[62460]: DEBUG nova.virt.hardware [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 626.962168] env[62460]: DEBUG nova.virt.hardware [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 626.962408] env[62460]: DEBUG nova.virt.hardware [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 626.962569] env[62460]: DEBUG nova.virt.hardware [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 626.962721] env[62460]: DEBUG nova.virt.hardware [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 626.962936] env[62460]: DEBUG nova.virt.hardware [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 626.963114] env[62460]: DEBUG nova.virt.hardware [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 626.963364] env[62460]: DEBUG nova.virt.hardware [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 626.963545] env[62460]: DEBUG nova.virt.hardware [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 626.963726] env[62460]: DEBUG nova.virt.hardware [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 626.964609] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-427aa727-f144-46e0-96f7-c2e197f921ea {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.975886] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c1f61de-80f1-4a52-91ae-36a4647edfb0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.008022] env[62460]: DEBUG nova.network.neutron [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.044858] env[62460]: DEBUG nova.compute.manager [None req-43eb215c-441e-47d9-a15c-59e321da1d3f tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] [instance: 3b71c366-cbd9-4b98-aa0d-c55b56e69231] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 627.047480] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c5b5d59-2d51-48e8-91a3-2e59e936169f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.171452] env[62460]: DEBUG nova.compute.manager [req-51aa3525-af5a-421e-9860-e04af503914e req-ad537390-85c5-4428-834b-6a33db0491aa service nova] [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] Received event network-changed-3f617afc-843a-4ab7-88fe-7a981af30004 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 627.173152] env[62460]: DEBUG nova.compute.manager [req-51aa3525-af5a-421e-9860-e04af503914e req-ad537390-85c5-4428-834b-6a33db0491aa service nova] [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] Refreshing instance network info cache due to event network-changed-3f617afc-843a-4ab7-88fe-7a981af30004. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 627.173152] env[62460]: DEBUG oslo_concurrency.lockutils [req-51aa3525-af5a-421e-9860-e04af503914e req-ad537390-85c5-4428-834b-6a33db0491aa service nova] Acquiring lock "refresh_cache-01bbc9ec-f2ee-4a48-a33c-784861e81097" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 627.173152] env[62460]: DEBUG oslo_concurrency.lockutils [req-51aa3525-af5a-421e-9860-e04af503914e req-ad537390-85c5-4428-834b-6a33db0491aa service nova] Acquired lock "refresh_cache-01bbc9ec-f2ee-4a48-a33c-784861e81097" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 627.173152] env[62460]: DEBUG nova.network.neutron [req-51aa3525-af5a-421e-9860-e04af503914e req-ad537390-85c5-4428-834b-6a33db0491aa service nova] [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] Refreshing network info cache for port 3f617afc-843a-4ab7-88fe-7a981af30004 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 627.279095] env[62460]: DEBUG nova.compute.manager [req-e28908c7-9065-44bc-8edc-92cb9083c23d req-0a0abb6f-aa34-4e4e-8ffb-c2e0a117ed7b service nova] [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] Received event network-vif-deleted-3373c77f-c40e-4658-9249-466f197b9940 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 627.314380] env[62460]: ERROR nova.compute.manager [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3f617afc-843a-4ab7-88fe-7a981af30004, please check neutron logs for more information. [ 627.314380] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 627.314380] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 627.314380] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 627.314380] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 627.314380] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 627.314380] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 627.314380] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 627.314380] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 627.314380] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 627.314380] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 627.314380] env[62460]: ERROR nova.compute.manager raise self.value [ 627.314380] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 627.314380] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 627.314380] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 627.314380] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 627.314982] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 627.314982] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 627.314982] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3f617afc-843a-4ab7-88fe-7a981af30004, please check neutron logs for more information. [ 627.314982] env[62460]: ERROR nova.compute.manager [ 627.314982] env[62460]: Traceback (most recent call last): [ 627.314982] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 627.314982] env[62460]: listener.cb(fileno) [ 627.314982] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 627.314982] env[62460]: result = function(*args, **kwargs) [ 627.314982] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 627.314982] env[62460]: return func(*args, **kwargs) [ 627.314982] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 627.314982] env[62460]: raise e [ 627.314982] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 627.314982] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 627.314982] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 627.314982] env[62460]: created_port_ids = self._update_ports_for_instance( [ 627.314982] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 627.314982] env[62460]: with excutils.save_and_reraise_exception(): [ 627.314982] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 627.314982] env[62460]: self.force_reraise() [ 627.314982] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 627.314982] env[62460]: raise self.value [ 627.314982] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 627.314982] env[62460]: updated_port = self._update_port( [ 627.314982] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 627.314982] env[62460]: _ensure_no_port_binding_failure(port) [ 627.314982] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 627.314982] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 627.315793] env[62460]: nova.exception.PortBindingFailed: Binding failed for port 3f617afc-843a-4ab7-88fe-7a981af30004, please check neutron logs for more information. [ 627.315793] env[62460]: Removing descriptor: 19 [ 627.315793] env[62460]: ERROR nova.compute.manager [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3f617afc-843a-4ab7-88fe-7a981af30004, please check neutron logs for more information. [ 627.315793] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] Traceback (most recent call last): [ 627.315793] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 627.315793] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] yield resources [ 627.315793] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 627.315793] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] self.driver.spawn(context, instance, image_meta, [ 627.315793] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 627.315793] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] self._vmops.spawn(context, instance, image_meta, injected_files, [ 627.315793] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 627.315793] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] vm_ref = self.build_virtual_machine(instance, [ 627.316149] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 627.316149] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] vif_infos = vmwarevif.get_vif_info(self._session, [ 627.316149] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 627.316149] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] for vif in network_info: [ 627.316149] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 627.316149] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] return self._sync_wrapper(fn, *args, **kwargs) [ 627.316149] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 627.316149] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] self.wait() [ 627.316149] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 627.316149] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] self[:] = self._gt.wait() [ 627.316149] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 627.316149] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] return self._exit_event.wait() [ 627.316149] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 627.316641] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] result = hub.switch() [ 627.316641] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 627.316641] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] return self.greenlet.switch() [ 627.316641] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 627.316641] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] result = function(*args, **kwargs) [ 627.316641] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 627.316641] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] return func(*args, **kwargs) [ 627.316641] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 627.316641] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] raise e [ 627.316641] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 627.316641] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] nwinfo = self.network_api.allocate_for_instance( [ 627.316641] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 627.316641] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] created_port_ids = self._update_ports_for_instance( [ 627.317025] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 627.317025] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] with excutils.save_and_reraise_exception(): [ 627.317025] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 627.317025] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] self.force_reraise() [ 627.317025] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 627.317025] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] raise self.value [ 627.317025] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 627.317025] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] updated_port = self._update_port( [ 627.317025] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 627.317025] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] _ensure_no_port_binding_failure(port) [ 627.317025] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 627.317025] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] raise exception.PortBindingFailed(port_id=port['id']) [ 627.317523] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] nova.exception.PortBindingFailed: Binding failed for port 3f617afc-843a-4ab7-88fe-7a981af30004, please check neutron logs for more information. [ 627.317523] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] [ 627.317523] env[62460]: INFO nova.compute.manager [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] Terminating instance [ 627.318345] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Acquiring lock "refresh_cache-01bbc9ec-f2ee-4a48-a33c-784861e81097" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 627.406437] env[62460]: DEBUG oslo_concurrency.lockutils [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 627.425858] env[62460]: DEBUG nova.network.neutron [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 627.471359] env[62460]: DEBUG nova.network.neutron [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.511396] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Releasing lock "refresh_cache-9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 627.511979] env[62460]: DEBUG nova.compute.manager [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 627.513374] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 627.513374] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a445de46-ce15-43f6-bb99-80a1db0ceb56 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.522577] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7385522-7c04-4a22-9a88-2076241bd45f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.545739] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6 could not be found. [ 627.545958] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 627.546165] env[62460]: INFO nova.compute.manager [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] Took 0.03 seconds to destroy the instance on the hypervisor. [ 627.546415] env[62460]: DEBUG oslo.service.loopingcall [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 627.546627] env[62460]: DEBUG nova.compute.manager [-] [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 627.546723] env[62460]: DEBUG nova.network.neutron [-] [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 627.556029] env[62460]: DEBUG oslo_concurrency.lockutils [None req-398cf1be-cafa-4d72-b54f-722129802cf2 tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Acquiring lock "3b71c366-cbd9-4b98-aa0d-c55b56e69231" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 627.556254] env[62460]: DEBUG oslo_concurrency.lockutils [None req-398cf1be-cafa-4d72-b54f-722129802cf2 tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Lock "3b71c366-cbd9-4b98-aa0d-c55b56e69231" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 627.556447] env[62460]: DEBUG oslo_concurrency.lockutils [None req-398cf1be-cafa-4d72-b54f-722129802cf2 tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Acquiring lock "3b71c366-cbd9-4b98-aa0d-c55b56e69231-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 627.556619] env[62460]: DEBUG oslo_concurrency.lockutils [None req-398cf1be-cafa-4d72-b54f-722129802cf2 tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Lock "3b71c366-cbd9-4b98-aa0d-c55b56e69231-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 627.556776] env[62460]: DEBUG oslo_concurrency.lockutils [None req-398cf1be-cafa-4d72-b54f-722129802cf2 tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Lock "3b71c366-cbd9-4b98-aa0d-c55b56e69231-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 627.559296] env[62460]: INFO nova.compute.manager [None req-398cf1be-cafa-4d72-b54f-722129802cf2 tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] [instance: 3b71c366-cbd9-4b98-aa0d-c55b56e69231] Terminating instance [ 627.560886] env[62460]: DEBUG oslo_concurrency.lockutils [None req-398cf1be-cafa-4d72-b54f-722129802cf2 tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Acquiring lock "refresh_cache-3b71c366-cbd9-4b98-aa0d-c55b56e69231" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 627.561046] env[62460]: DEBUG oslo_concurrency.lockutils [None req-398cf1be-cafa-4d72-b54f-722129802cf2 tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Acquired lock "refresh_cache-3b71c366-cbd9-4b98-aa0d-c55b56e69231" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 627.561263] env[62460]: DEBUG nova.network.neutron [None req-398cf1be-cafa-4d72-b54f-722129802cf2 tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] [instance: 3b71c366-cbd9-4b98-aa0d-c55b56e69231] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 627.562594] env[62460]: INFO nova.compute.manager [None req-43eb215c-441e-47d9-a15c-59e321da1d3f tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] [instance: 3b71c366-cbd9-4b98-aa0d-c55b56e69231] instance snapshotting [ 627.563040] env[62460]: DEBUG nova.objects.instance [None req-43eb215c-441e-47d9-a15c-59e321da1d3f tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Lazy-loading 'flavor' on Instance uuid 3b71c366-cbd9-4b98-aa0d-c55b56e69231 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 627.564649] env[62460]: DEBUG nova.network.neutron [-] [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 627.696988] env[62460]: DEBUG nova.network.neutron [req-51aa3525-af5a-421e-9860-e04af503914e req-ad537390-85c5-4428-834b-6a33db0491aa service nova] [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 627.885450] env[62460]: DEBUG nova.network.neutron [req-51aa3525-af5a-421e-9860-e04af503914e req-ad537390-85c5-4428-834b-6a33db0491aa service nova] [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.973347] env[62460]: DEBUG oslo_concurrency.lockutils [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Releasing lock "refresh_cache-443e235d-32b2-4af7-bdae-3f30e8196ee3" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 627.973592] env[62460]: DEBUG nova.compute.manager [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 627.973759] env[62460]: DEBUG nova.compute.manager [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 627.973930] env[62460]: DEBUG nova.network.neutron [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 627.995160] env[62460]: DEBUG nova.network.neutron [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 628.074302] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-014c4026-c211-4c66-b48a-1c3462cc67f9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.081028] env[62460]: DEBUG nova.network.neutron [-] [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.103507] env[62460]: DEBUG nova.network.neutron [None req-398cf1be-cafa-4d72-b54f-722129802cf2 tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] [instance: 3b71c366-cbd9-4b98-aa0d-c55b56e69231] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 628.106371] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d0e99cd-3145-43cc-9294-d427b482c662 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.226653] env[62460]: DEBUG nova.network.neutron [None req-398cf1be-cafa-4d72-b54f-722129802cf2 tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] [instance: 3b71c366-cbd9-4b98-aa0d-c55b56e69231] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.305273] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a070fb29-cfeb-48c3-95a9-f01e43e94149 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.313210] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcb11be6-a620-4646-ae78-c08ba86d8498 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.344642] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4922de0b-54d1-48f2-a409-cf7b765499ad {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.352477] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cd50701-4c0d-4d83-96b4-b69c27ad0f8c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.367026] env[62460]: DEBUG nova.compute.provider_tree [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 628.388574] env[62460]: DEBUG oslo_concurrency.lockutils [req-51aa3525-af5a-421e-9860-e04af503914e req-ad537390-85c5-4428-834b-6a33db0491aa service nova] Releasing lock "refresh_cache-01bbc9ec-f2ee-4a48-a33c-784861e81097" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 628.388574] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Acquired lock "refresh_cache-01bbc9ec-f2ee-4a48-a33c-784861e81097" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 628.388574] env[62460]: DEBUG nova.network.neutron [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 628.498023] env[62460]: DEBUG nova.network.neutron [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.585262] env[62460]: INFO nova.compute.manager [-] [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] Took 1.04 seconds to deallocate network for instance. [ 628.586603] env[62460]: DEBUG nova.compute.claims [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 628.586971] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 628.619197] env[62460]: DEBUG nova.compute.manager [None req-43eb215c-441e-47d9-a15c-59e321da1d3f tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] [instance: 3b71c366-cbd9-4b98-aa0d-c55b56e69231] Instance disappeared during snapshot {{(pid=62460) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 628.732023] env[62460]: DEBUG oslo_concurrency.lockutils [None req-398cf1be-cafa-4d72-b54f-722129802cf2 tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Releasing lock "refresh_cache-3b71c366-cbd9-4b98-aa0d-c55b56e69231" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 628.733017] env[62460]: DEBUG nova.compute.manager [None req-398cf1be-cafa-4d72-b54f-722129802cf2 tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] [instance: 3b71c366-cbd9-4b98-aa0d-c55b56e69231] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 628.733017] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-398cf1be-cafa-4d72-b54f-722129802cf2 tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] [instance: 3b71c366-cbd9-4b98-aa0d-c55b56e69231] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 628.733780] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7fbf15a-a281-4bdf-a17d-ab7b690e4a10 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.748120] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-398cf1be-cafa-4d72-b54f-722129802cf2 tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] [instance: 3b71c366-cbd9-4b98-aa0d-c55b56e69231] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 628.751023] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9c00380a-4fa4-40cb-b7e2-f232bedfe220 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.755445] env[62460]: DEBUG oslo_vmware.api [None req-398cf1be-cafa-4d72-b54f-722129802cf2 tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Waiting for the task: (returnval){ [ 628.755445] env[62460]: value = "task-1313446" [ 628.755445] env[62460]: _type = "Task" [ 628.755445] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.764781] env[62460]: DEBUG oslo_vmware.api [None req-398cf1be-cafa-4d72-b54f-722129802cf2 tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Task: {'id': task-1313446, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.793071] env[62460]: DEBUG nova.compute.manager [None req-43eb215c-441e-47d9-a15c-59e321da1d3f tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] [instance: 3b71c366-cbd9-4b98-aa0d-c55b56e69231] Found 0 images (rotation: 2) {{(pid=62460) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 628.871472] env[62460]: DEBUG nova.scheduler.client.report [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 628.912411] env[62460]: DEBUG nova.network.neutron [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 629.000155] env[62460]: INFO nova.compute.manager [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] [instance: 443e235d-32b2-4af7-bdae-3f30e8196ee3] Took 1.03 seconds to deallocate network for instance. [ 629.019582] env[62460]: DEBUG nova.network.neutron [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.201811] env[62460]: DEBUG nova.compute.manager [req-d81ae2c5-a0d2-4460-a817-0ee90aa61dbf req-e35f9a96-3944-4c31-adae-5545999525bd service nova] [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] Received event network-vif-deleted-3f617afc-843a-4ab7-88fe-7a981af30004 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 629.270038] env[62460]: DEBUG oslo_vmware.api [None req-398cf1be-cafa-4d72-b54f-722129802cf2 tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Task: {'id': task-1313446, 'name': PowerOffVM_Task, 'duration_secs': 0.161667} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.270038] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-398cf1be-cafa-4d72-b54f-722129802cf2 tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] [instance: 3b71c366-cbd9-4b98-aa0d-c55b56e69231] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 629.270038] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-398cf1be-cafa-4d72-b54f-722129802cf2 tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] [instance: 3b71c366-cbd9-4b98-aa0d-c55b56e69231] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 629.270038] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4fdc86b7-1231-492a-9f07-93580ab971f8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.295191] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-398cf1be-cafa-4d72-b54f-722129802cf2 tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] [instance: 3b71c366-cbd9-4b98-aa0d-c55b56e69231] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 629.295440] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-398cf1be-cafa-4d72-b54f-722129802cf2 tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] [instance: 3b71c366-cbd9-4b98-aa0d-c55b56e69231] Deleting contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 629.295627] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-398cf1be-cafa-4d72-b54f-722129802cf2 tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Deleting the datastore file [datastore1] 3b71c366-cbd9-4b98-aa0d-c55b56e69231 {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 629.296158] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ea5e2627-75d8-4e53-a48a-2e50fadaacdc {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.303078] env[62460]: DEBUG oslo_vmware.api [None req-398cf1be-cafa-4d72-b54f-722129802cf2 tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Waiting for the task: (returnval){ [ 629.303078] env[62460]: value = "task-1313448" [ 629.303078] env[62460]: _type = "Task" [ 629.303078] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.312250] env[62460]: DEBUG oslo_vmware.api [None req-398cf1be-cafa-4d72-b54f-722129802cf2 tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Task: {'id': task-1313448, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.381427] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.481s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 629.381973] env[62460]: DEBUG nova.compute.manager [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] [instance: 94703126-bb31-40af-b945-2ef5ff37e094] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 629.384605] env[62460]: DEBUG oslo_concurrency.lockutils [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.059s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 629.386111] env[62460]: INFO nova.compute.claims [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 629.524566] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Releasing lock "refresh_cache-01bbc9ec-f2ee-4a48-a33c-784861e81097" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 629.524843] env[62460]: DEBUG nova.compute.manager [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 629.524898] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 629.525220] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ee81fb4d-ec69-488b-a177-7e9f68be4483 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.535031] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a51577fe-0d60-4f0d-b94b-aff23dba291f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.556387] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 01bbc9ec-f2ee-4a48-a33c-784861e81097 could not be found. [ 629.556627] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 629.556817] env[62460]: INFO nova.compute.manager [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] Took 0.03 seconds to destroy the instance on the hypervisor. [ 629.557132] env[62460]: DEBUG oslo.service.loopingcall [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 629.557368] env[62460]: DEBUG nova.compute.manager [-] [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 629.557464] env[62460]: DEBUG nova.network.neutron [-] [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 629.576534] env[62460]: DEBUG nova.network.neutron [-] [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 629.824153] env[62460]: DEBUG oslo_vmware.api [None req-398cf1be-cafa-4d72-b54f-722129802cf2 tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Task: {'id': task-1313448, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.106395} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.824153] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-398cf1be-cafa-4d72-b54f-722129802cf2 tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 629.824153] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-398cf1be-cafa-4d72-b54f-722129802cf2 tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] [instance: 3b71c366-cbd9-4b98-aa0d-c55b56e69231] Deleted contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 629.824153] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-398cf1be-cafa-4d72-b54f-722129802cf2 tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] [instance: 3b71c366-cbd9-4b98-aa0d-c55b56e69231] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 629.824153] env[62460]: INFO nova.compute.manager [None req-398cf1be-cafa-4d72-b54f-722129802cf2 tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] [instance: 3b71c366-cbd9-4b98-aa0d-c55b56e69231] Took 1.09 seconds to destroy the instance on the hypervisor. [ 629.824307] env[62460]: DEBUG oslo.service.loopingcall [None req-398cf1be-cafa-4d72-b54f-722129802cf2 tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 629.824307] env[62460]: DEBUG nova.compute.manager [-] [instance: 3b71c366-cbd9-4b98-aa0d-c55b56e69231] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 629.824307] env[62460]: DEBUG nova.network.neutron [-] [instance: 3b71c366-cbd9-4b98-aa0d-c55b56e69231] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 629.840043] env[62460]: DEBUG nova.network.neutron [-] [instance: 3b71c366-cbd9-4b98-aa0d-c55b56e69231] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 629.893291] env[62460]: DEBUG nova.compute.utils [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 629.896696] env[62460]: DEBUG nova.compute.manager [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] [instance: 94703126-bb31-40af-b945-2ef5ff37e094] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 629.896868] env[62460]: DEBUG nova.network.neutron [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] [instance: 94703126-bb31-40af-b945-2ef5ff37e094] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 629.975726] env[62460]: DEBUG nova.policy [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b20cb1a3df8b43a0b65883f054aff0b0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f99d2add2f7140abb255cbe141a78be6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 630.058879] env[62460]: INFO nova.scheduler.client.report [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Deleted allocations for instance 443e235d-32b2-4af7-bdae-3f30e8196ee3 [ 630.082026] env[62460]: DEBUG nova.network.neutron [-] [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.343041] env[62460]: DEBUG nova.network.neutron [-] [instance: 3b71c366-cbd9-4b98-aa0d-c55b56e69231] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.400164] env[62460]: DEBUG nova.compute.manager [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] [instance: 94703126-bb31-40af-b945-2ef5ff37e094] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 630.495286] env[62460]: DEBUG nova.network.neutron [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] [instance: 94703126-bb31-40af-b945-2ef5ff37e094] Successfully created port: 620a0798-9892-49ff-9a8c-a60a330f5b35 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 630.568417] env[62460]: DEBUG oslo_concurrency.lockutils [None req-be617c24-8359-4849-8eb8-bf231fc11561 tempest-DeleteServersAdminTestJSON-538413674 tempest-DeleteServersAdminTestJSON-538413674-project-member] Lock "443e235d-32b2-4af7-bdae-3f30e8196ee3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 75.062s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 630.568702] env[62460]: DEBUG oslo_concurrency.lockutils [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] Acquiring lock "c4c5abc4-a5ed-4337-a930-d37d579819f8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 630.568805] env[62460]: DEBUG oslo_concurrency.lockutils [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] Lock "c4c5abc4-a5ed-4337-a930-d37d579819f8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.587220] env[62460]: INFO nova.compute.manager [-] [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] Took 1.03 seconds to deallocate network for instance. [ 630.591606] env[62460]: DEBUG nova.compute.claims [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 630.591606] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 630.776220] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ed034b9-cc9f-4427-ab45-4131e0f7d0df {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.784360] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d5aba77-e15e-4740-a270-b603b072c15a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.814607] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78dcf244-e2d6-49df-bbbb-d4fec93f2702 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.824762] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b32ff79-4728-46f0-85ff-6a97aa203a8c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.838430] env[62460]: DEBUG nova.compute.provider_tree [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 630.848739] env[62460]: INFO nova.compute.manager [-] [instance: 3b71c366-cbd9-4b98-aa0d-c55b56e69231] Took 1.03 seconds to deallocate network for instance. [ 631.072285] env[62460]: DEBUG nova.compute.manager [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 631.342314] env[62460]: DEBUG nova.scheduler.client.report [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 631.356163] env[62460]: DEBUG oslo_concurrency.lockutils [None req-398cf1be-cafa-4d72-b54f-722129802cf2 tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 631.414771] env[62460]: DEBUG nova.compute.manager [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] [instance: 94703126-bb31-40af-b945-2ef5ff37e094] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 631.450246] env[62460]: DEBUG nova.virt.hardware [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 631.450492] env[62460]: DEBUG nova.virt.hardware [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 631.450653] env[62460]: DEBUG nova.virt.hardware [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 631.450834] env[62460]: DEBUG nova.virt.hardware [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 631.454251] env[62460]: DEBUG nova.virt.hardware [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 631.454505] env[62460]: DEBUG nova.virt.hardware [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 631.454752] env[62460]: DEBUG nova.virt.hardware [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 631.454894] env[62460]: DEBUG nova.virt.hardware [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 631.455090] env[62460]: DEBUG nova.virt.hardware [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 631.455260] env[62460]: DEBUG nova.virt.hardware [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 631.455436] env[62460]: DEBUG nova.virt.hardware [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 631.457389] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78ba5c28-94e2-47e2-aa13-16c212afff0a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.464616] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5a9ee3f-361a-4c33-95fe-e79c2a865e0a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.610428] env[62460]: DEBUG oslo_concurrency.lockutils [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 631.732315] env[62460]: DEBUG nova.compute.manager [req-f962dd11-46e6-4c81-bf96-8beedb51d219 req-8a759124-c2db-41d9-a195-8c2e7dcdbdb8 service nova] [instance: 94703126-bb31-40af-b945-2ef5ff37e094] Received event network-changed-620a0798-9892-49ff-9a8c-a60a330f5b35 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 631.732315] env[62460]: DEBUG nova.compute.manager [req-f962dd11-46e6-4c81-bf96-8beedb51d219 req-8a759124-c2db-41d9-a195-8c2e7dcdbdb8 service nova] [instance: 94703126-bb31-40af-b945-2ef5ff37e094] Refreshing instance network info cache due to event network-changed-620a0798-9892-49ff-9a8c-a60a330f5b35. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 631.732315] env[62460]: DEBUG oslo_concurrency.lockutils [req-f962dd11-46e6-4c81-bf96-8beedb51d219 req-8a759124-c2db-41d9-a195-8c2e7dcdbdb8 service nova] Acquiring lock "refresh_cache-94703126-bb31-40af-b945-2ef5ff37e094" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 631.732315] env[62460]: DEBUG oslo_concurrency.lockutils [req-f962dd11-46e6-4c81-bf96-8beedb51d219 req-8a759124-c2db-41d9-a195-8c2e7dcdbdb8 service nova] Acquired lock "refresh_cache-94703126-bb31-40af-b945-2ef5ff37e094" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.732315] env[62460]: DEBUG nova.network.neutron [req-f962dd11-46e6-4c81-bf96-8beedb51d219 req-8a759124-c2db-41d9-a195-8c2e7dcdbdb8 service nova] [instance: 94703126-bb31-40af-b945-2ef5ff37e094] Refreshing network info cache for port 620a0798-9892-49ff-9a8c-a60a330f5b35 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 631.854351] env[62460]: DEBUG oslo_concurrency.lockutils [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.468s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 631.854351] env[62460]: DEBUG nova.compute.manager [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 631.857086] env[62460]: DEBUG oslo_concurrency.lockutils [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.532s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 631.858973] env[62460]: INFO nova.compute.claims [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 631.896480] env[62460]: ERROR nova.compute.manager [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 620a0798-9892-49ff-9a8c-a60a330f5b35, please check neutron logs for more information. [ 631.896480] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 631.896480] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 631.896480] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 631.896480] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 631.896480] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 631.896480] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 631.896480] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 631.896480] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.896480] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 631.896480] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.896480] env[62460]: ERROR nova.compute.manager raise self.value [ 631.896480] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 631.896480] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 631.896480] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.896480] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 631.896885] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.896885] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 631.896885] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 620a0798-9892-49ff-9a8c-a60a330f5b35, please check neutron logs for more information. [ 631.896885] env[62460]: ERROR nova.compute.manager [ 631.896885] env[62460]: Traceback (most recent call last): [ 631.896885] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 631.896885] env[62460]: listener.cb(fileno) [ 631.896885] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 631.896885] env[62460]: result = function(*args, **kwargs) [ 631.896885] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 631.896885] env[62460]: return func(*args, **kwargs) [ 631.896885] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 631.896885] env[62460]: raise e [ 631.896885] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 631.896885] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 631.896885] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 631.896885] env[62460]: created_port_ids = self._update_ports_for_instance( [ 631.896885] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 631.896885] env[62460]: with excutils.save_and_reraise_exception(): [ 631.896885] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.896885] env[62460]: self.force_reraise() [ 631.896885] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.896885] env[62460]: raise self.value [ 631.896885] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 631.896885] env[62460]: updated_port = self._update_port( [ 631.896885] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.896885] env[62460]: _ensure_no_port_binding_failure(port) [ 631.896885] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.896885] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 631.897640] env[62460]: nova.exception.PortBindingFailed: Binding failed for port 620a0798-9892-49ff-9a8c-a60a330f5b35, please check neutron logs for more information. [ 631.897640] env[62460]: Removing descriptor: 19 [ 631.897640] env[62460]: ERROR nova.compute.manager [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] [instance: 94703126-bb31-40af-b945-2ef5ff37e094] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 620a0798-9892-49ff-9a8c-a60a330f5b35, please check neutron logs for more information. [ 631.897640] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] Traceback (most recent call last): [ 631.897640] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 631.897640] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] yield resources [ 631.897640] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 631.897640] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] self.driver.spawn(context, instance, image_meta, [ 631.897640] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 631.897640] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] self._vmops.spawn(context, instance, image_meta, injected_files, [ 631.897640] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 631.897640] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] vm_ref = self.build_virtual_machine(instance, [ 631.897938] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 631.897938] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] vif_infos = vmwarevif.get_vif_info(self._session, [ 631.897938] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 631.897938] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] for vif in network_info: [ 631.897938] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 631.897938] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] return self._sync_wrapper(fn, *args, **kwargs) [ 631.897938] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 631.897938] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] self.wait() [ 631.897938] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 631.897938] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] self[:] = self._gt.wait() [ 631.897938] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 631.897938] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] return self._exit_event.wait() [ 631.897938] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 631.898699] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] result = hub.switch() [ 631.898699] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 631.898699] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] return self.greenlet.switch() [ 631.898699] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 631.898699] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] result = function(*args, **kwargs) [ 631.898699] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 631.898699] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] return func(*args, **kwargs) [ 631.898699] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 631.898699] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] raise e [ 631.898699] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 631.898699] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] nwinfo = self.network_api.allocate_for_instance( [ 631.898699] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 631.898699] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] created_port_ids = self._update_ports_for_instance( [ 631.899103] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 631.899103] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] with excutils.save_and_reraise_exception(): [ 631.899103] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.899103] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] self.force_reraise() [ 631.899103] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.899103] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] raise self.value [ 631.899103] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 631.899103] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] updated_port = self._update_port( [ 631.899103] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.899103] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] _ensure_no_port_binding_failure(port) [ 631.899103] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.899103] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] raise exception.PortBindingFailed(port_id=port['id']) [ 631.899393] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] nova.exception.PortBindingFailed: Binding failed for port 620a0798-9892-49ff-9a8c-a60a330f5b35, please check neutron logs for more information. [ 631.899393] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] [ 631.899393] env[62460]: INFO nova.compute.manager [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] [instance: 94703126-bb31-40af-b945-2ef5ff37e094] Terminating instance [ 631.901053] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] Acquiring lock "refresh_cache-94703126-bb31-40af-b945-2ef5ff37e094" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 632.248482] env[62460]: DEBUG nova.network.neutron [req-f962dd11-46e6-4c81-bf96-8beedb51d219 req-8a759124-c2db-41d9-a195-8c2e7dcdbdb8 service nova] [instance: 94703126-bb31-40af-b945-2ef5ff37e094] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 632.311805] env[62460]: DEBUG nova.network.neutron [req-f962dd11-46e6-4c81-bf96-8beedb51d219 req-8a759124-c2db-41d9-a195-8c2e7dcdbdb8 service nova] [instance: 94703126-bb31-40af-b945-2ef5ff37e094] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.364387] env[62460]: DEBUG nova.compute.utils [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 632.367752] env[62460]: DEBUG nova.compute.manager [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 632.367936] env[62460]: DEBUG nova.network.neutron [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 632.404639] env[62460]: DEBUG nova.policy [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0bd87e1bf37a4c57966fa68415784943', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '97bb592cba2e46ceb02e311a754c9860', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 632.768815] env[62460]: DEBUG nova.network.neutron [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] Successfully created port: 10776c82-fe5d-46e6-b9e6-7aacb6b11ab9 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 632.814117] env[62460]: DEBUG oslo_concurrency.lockutils [req-f962dd11-46e6-4c81-bf96-8beedb51d219 req-8a759124-c2db-41d9-a195-8c2e7dcdbdb8 service nova] Releasing lock "refresh_cache-94703126-bb31-40af-b945-2ef5ff37e094" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 632.814568] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] Acquired lock "refresh_cache-94703126-bb31-40af-b945-2ef5ff37e094" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 632.814760] env[62460]: DEBUG nova.network.neutron [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] [instance: 94703126-bb31-40af-b945-2ef5ff37e094] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 632.876174] env[62460]: DEBUG nova.compute.manager [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 633.163014] env[62460]: DEBUG oslo_concurrency.lockutils [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Acquiring lock "4ed90d16-81a6-4dbd-8936-0e137151171f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.166321] env[62460]: DEBUG oslo_concurrency.lockutils [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Lock "4ed90d16-81a6-4dbd-8936-0e137151171f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.003s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 633.212020] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a097e0f-a03d-47f1-a0f4-ed730750b01c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.220402] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c50634e6-2245-45ee-82e1-a2c70c36e1a5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.254181] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-897ba416-aabb-46e5-9428-cfe823561a16 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.262209] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b749b2f-6e1e-461c-8a27-db04ae144305 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.276427] env[62460]: DEBUG nova.compute.provider_tree [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 633.332600] env[62460]: DEBUG nova.network.neutron [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] [instance: 94703126-bb31-40af-b945-2ef5ff37e094] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 633.393454] env[62460]: DEBUG nova.network.neutron [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] [instance: 94703126-bb31-40af-b945-2ef5ff37e094] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.434271] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Acquiring lock "5214f4a1-3e28-41bf-88d9-161511385e1b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.434599] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Lock "5214f4a1-3e28-41bf-88d9-161511385e1b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 633.775921] env[62460]: ERROR nova.compute.manager [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 10776c82-fe5d-46e6-b9e6-7aacb6b11ab9, please check neutron logs for more information. [ 633.775921] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 633.775921] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 633.775921] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 633.775921] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 633.775921] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 633.775921] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 633.775921] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 633.775921] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 633.775921] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 633.775921] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 633.775921] env[62460]: ERROR nova.compute.manager raise self.value [ 633.775921] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 633.775921] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 633.775921] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 633.775921] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 633.776596] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 633.776596] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 633.776596] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 10776c82-fe5d-46e6-b9e6-7aacb6b11ab9, please check neutron logs for more information. [ 633.776596] env[62460]: ERROR nova.compute.manager [ 633.776596] env[62460]: Traceback (most recent call last): [ 633.776596] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 633.776596] env[62460]: listener.cb(fileno) [ 633.776596] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 633.776596] env[62460]: result = function(*args, **kwargs) [ 633.776596] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 633.776596] env[62460]: return func(*args, **kwargs) [ 633.776596] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 633.776596] env[62460]: raise e [ 633.776596] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 633.776596] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 633.776596] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 633.776596] env[62460]: created_port_ids = self._update_ports_for_instance( [ 633.776596] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 633.776596] env[62460]: with excutils.save_and_reraise_exception(): [ 633.776596] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 633.776596] env[62460]: self.force_reraise() [ 633.776596] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 633.776596] env[62460]: raise self.value [ 633.776596] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 633.776596] env[62460]: updated_port = self._update_port( [ 633.776596] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 633.776596] env[62460]: _ensure_no_port_binding_failure(port) [ 633.776596] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 633.776596] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 633.777580] env[62460]: nova.exception.PortBindingFailed: Binding failed for port 10776c82-fe5d-46e6-b9e6-7aacb6b11ab9, please check neutron logs for more information. [ 633.777580] env[62460]: Removing descriptor: 19 [ 633.780029] env[62460]: DEBUG nova.scheduler.client.report [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 633.804666] env[62460]: DEBUG nova.compute.manager [req-9463570f-9e00-47cd-90b5-e120ca8bd8a4 req-0807f7bf-04b1-4464-a220-bbc1507636ae service nova] [instance: 94703126-bb31-40af-b945-2ef5ff37e094] Received event network-vif-deleted-620a0798-9892-49ff-9a8c-a60a330f5b35 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 633.804925] env[62460]: DEBUG nova.compute.manager [req-9463570f-9e00-47cd-90b5-e120ca8bd8a4 req-0807f7bf-04b1-4464-a220-bbc1507636ae service nova] [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] Received event network-changed-10776c82-fe5d-46e6-b9e6-7aacb6b11ab9 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 633.805143] env[62460]: DEBUG nova.compute.manager [req-9463570f-9e00-47cd-90b5-e120ca8bd8a4 req-0807f7bf-04b1-4464-a220-bbc1507636ae service nova] [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] Refreshing instance network info cache due to event network-changed-10776c82-fe5d-46e6-b9e6-7aacb6b11ab9. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 633.805390] env[62460]: DEBUG oslo_concurrency.lockutils [req-9463570f-9e00-47cd-90b5-e120ca8bd8a4 req-0807f7bf-04b1-4464-a220-bbc1507636ae service nova] Acquiring lock "refresh_cache-3426ace3-0cff-4119-85fe-31e681a8597b" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 633.805566] env[62460]: DEBUG oslo_concurrency.lockutils [req-9463570f-9e00-47cd-90b5-e120ca8bd8a4 req-0807f7bf-04b1-4464-a220-bbc1507636ae service nova] Acquired lock "refresh_cache-3426ace3-0cff-4119-85fe-31e681a8597b" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.805764] env[62460]: DEBUG nova.network.neutron [req-9463570f-9e00-47cd-90b5-e120ca8bd8a4 req-0807f7bf-04b1-4464-a220-bbc1507636ae service nova] [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] Refreshing network info cache for port 10776c82-fe5d-46e6-b9e6-7aacb6b11ab9 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 633.886580] env[62460]: DEBUG nova.compute.manager [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 633.898406] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] Releasing lock "refresh_cache-94703126-bb31-40af-b945-2ef5ff37e094" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 633.898490] env[62460]: DEBUG nova.compute.manager [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] [instance: 94703126-bb31-40af-b945-2ef5ff37e094] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 633.898681] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] [instance: 94703126-bb31-40af-b945-2ef5ff37e094] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 633.898960] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0b66fcd5-5688-4f7d-a0d3-5e10f9dbae8a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.909058] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98cf48ae-6749-44a2-9cfe-5ba6c492f9d2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.920783] env[62460]: DEBUG nova.virt.hardware [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 633.921038] env[62460]: DEBUG nova.virt.hardware [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 633.921295] env[62460]: DEBUG nova.virt.hardware [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 633.921501] env[62460]: DEBUG nova.virt.hardware [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 633.921653] env[62460]: DEBUG nova.virt.hardware [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 633.921804] env[62460]: DEBUG nova.virt.hardware [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 633.922027] env[62460]: DEBUG nova.virt.hardware [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 633.922202] env[62460]: DEBUG nova.virt.hardware [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 633.922397] env[62460]: DEBUG nova.virt.hardware [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 633.922568] env[62460]: DEBUG nova.virt.hardware [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 633.922742] env[62460]: DEBUG nova.virt.hardware [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 633.923492] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f8dd66b-0276-475d-871c-ccd39592fe9e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.930877] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebda1a14-4eba-4d4f-aa14-e160ac6d1026 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.938367] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] [instance: 94703126-bb31-40af-b945-2ef5ff37e094] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 94703126-bb31-40af-b945-2ef5ff37e094 could not be found. [ 633.938577] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] [instance: 94703126-bb31-40af-b945-2ef5ff37e094] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 633.938761] env[62460]: INFO nova.compute.manager [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] [instance: 94703126-bb31-40af-b945-2ef5ff37e094] Took 0.04 seconds to destroy the instance on the hypervisor. [ 633.939007] env[62460]: DEBUG oslo.service.loopingcall [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 633.939562] env[62460]: DEBUG nova.compute.manager [-] [instance: 94703126-bb31-40af-b945-2ef5ff37e094] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 633.940091] env[62460]: DEBUG nova.network.neutron [-] [instance: 94703126-bb31-40af-b945-2ef5ff37e094] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 633.949970] env[62460]: ERROR nova.compute.manager [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 10776c82-fe5d-46e6-b9e6-7aacb6b11ab9, please check neutron logs for more information. [ 633.949970] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] Traceback (most recent call last): [ 633.949970] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 633.949970] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] yield resources [ 633.949970] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 633.949970] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] self.driver.spawn(context, instance, image_meta, [ 633.949970] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 633.949970] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 633.949970] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 633.949970] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] vm_ref = self.build_virtual_machine(instance, [ 633.949970] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 633.950450] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] vif_infos = vmwarevif.get_vif_info(self._session, [ 633.950450] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 633.950450] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] for vif in network_info: [ 633.950450] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 633.950450] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] return self._sync_wrapper(fn, *args, **kwargs) [ 633.950450] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 633.950450] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] self.wait() [ 633.950450] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 633.950450] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] self[:] = self._gt.wait() [ 633.950450] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 633.950450] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] return self._exit_event.wait() [ 633.950450] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 633.950450] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] current.throw(*self._exc) [ 633.950769] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 633.950769] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] result = function(*args, **kwargs) [ 633.950769] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 633.950769] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] return func(*args, **kwargs) [ 633.950769] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 633.950769] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] raise e [ 633.950769] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 633.950769] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] nwinfo = self.network_api.allocate_for_instance( [ 633.950769] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 633.950769] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] created_port_ids = self._update_ports_for_instance( [ 633.950769] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 633.950769] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] with excutils.save_and_reraise_exception(): [ 633.950769] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 633.951090] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] self.force_reraise() [ 633.951090] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 633.951090] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] raise self.value [ 633.951090] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 633.951090] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] updated_port = self._update_port( [ 633.951090] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 633.951090] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] _ensure_no_port_binding_failure(port) [ 633.951090] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 633.951090] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] raise exception.PortBindingFailed(port_id=port['id']) [ 633.951090] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] nova.exception.PortBindingFailed: Binding failed for port 10776c82-fe5d-46e6-b9e6-7aacb6b11ab9, please check neutron logs for more information. [ 633.951090] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] [ 633.951090] env[62460]: INFO nova.compute.manager [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] Terminating instance [ 633.952170] env[62460]: DEBUG oslo_concurrency.lockutils [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] Acquiring lock "refresh_cache-3426ace3-0cff-4119-85fe-31e681a8597b" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 633.973861] env[62460]: DEBUG nova.network.neutron [-] [instance: 94703126-bb31-40af-b945-2ef5ff37e094] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 634.285917] env[62460]: DEBUG oslo_concurrency.lockutils [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.429s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 634.287139] env[62460]: DEBUG nova.compute.manager [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 634.289855] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 16.269s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 634.328309] env[62460]: DEBUG nova.network.neutron [req-9463570f-9e00-47cd-90b5-e120ca8bd8a4 req-0807f7bf-04b1-4464-a220-bbc1507636ae service nova] [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 634.408507] env[62460]: DEBUG nova.network.neutron [req-9463570f-9e00-47cd-90b5-e120ca8bd8a4 req-0807f7bf-04b1-4464-a220-bbc1507636ae service nova] [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.476406] env[62460]: DEBUG nova.network.neutron [-] [instance: 94703126-bb31-40af-b945-2ef5ff37e094] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.792469] env[62460]: DEBUG nova.compute.utils [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 634.797306] env[62460]: DEBUG nova.compute.manager [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 634.797484] env[62460]: DEBUG nova.network.neutron [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 634.853679] env[62460]: DEBUG nova.policy [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5d0676367f10488e9aa21fa9f64bbb2f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '91a65d30ffe14c9f8c5d869164ce7b86', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 634.911395] env[62460]: DEBUG oslo_concurrency.lockutils [req-9463570f-9e00-47cd-90b5-e120ca8bd8a4 req-0807f7bf-04b1-4464-a220-bbc1507636ae service nova] Releasing lock "refresh_cache-3426ace3-0cff-4119-85fe-31e681a8597b" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 634.911830] env[62460]: DEBUG oslo_concurrency.lockutils [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] Acquired lock "refresh_cache-3426ace3-0cff-4119-85fe-31e681a8597b" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 634.912039] env[62460]: DEBUG nova.network.neutron [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 634.979104] env[62460]: INFO nova.compute.manager [-] [instance: 94703126-bb31-40af-b945-2ef5ff37e094] Took 1.04 seconds to deallocate network for instance. [ 634.981435] env[62460]: DEBUG nova.compute.claims [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] [instance: 94703126-bb31-40af-b945-2ef5ff37e094] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 634.981614] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 635.120781] env[62460]: DEBUG nova.network.neutron [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] Successfully created port: 3221cf96-ce6e-40d8-bb3b-7ed671e0b4d1 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 635.298183] env[62460]: DEBUG nova.compute.manager [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 635.333019] env[62460]: WARNING nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 3b71c366-cbd9-4b98-aa0d-c55b56e69231 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 635.333019] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 635.333019] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 01bbc9ec-f2ee-4a48-a33c-784861e81097 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 635.333019] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 94703126-bb31-40af-b945-2ef5ff37e094 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 635.333199] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 3426ace3-0cff-4119-85fe-31e681a8597b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 635.333199] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance dba18c87-50a2-4ac8-8bd2-77f8f62b723c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 635.438598] env[62460]: DEBUG nova.network.neutron [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 635.528491] env[62460]: DEBUG nova.network.neutron [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 635.836478] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 01a3e1be-d29b-45d7-987e-66a4395ae2a4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 635.857581] env[62460]: DEBUG nova.compute.manager [req-99cda7e9-01d2-483a-821f-e0825e25bd62 req-7876d61b-159f-426d-b35e-f6f033e581c2 service nova] [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] Received event network-vif-deleted-10776c82-fe5d-46e6-b9e6-7aacb6b11ab9 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 636.031958] env[62460]: DEBUG oslo_concurrency.lockutils [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] Releasing lock "refresh_cache-3426ace3-0cff-4119-85fe-31e681a8597b" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 636.032483] env[62460]: DEBUG nova.compute.manager [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 636.032699] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 636.033919] env[62460]: ERROR nova.compute.manager [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3221cf96-ce6e-40d8-bb3b-7ed671e0b4d1, please check neutron logs for more information. [ 636.033919] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 636.033919] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 636.033919] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 636.033919] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 636.033919] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 636.033919] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 636.033919] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 636.033919] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 636.033919] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 636.033919] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 636.033919] env[62460]: ERROR nova.compute.manager raise self.value [ 636.033919] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 636.033919] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 636.033919] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 636.033919] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 636.034306] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 636.034306] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 636.034306] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3221cf96-ce6e-40d8-bb3b-7ed671e0b4d1, please check neutron logs for more information. [ 636.034306] env[62460]: ERROR nova.compute.manager [ 636.034306] env[62460]: Traceback (most recent call last): [ 636.034306] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 636.034306] env[62460]: listener.cb(fileno) [ 636.034306] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 636.034306] env[62460]: result = function(*args, **kwargs) [ 636.034306] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 636.034306] env[62460]: return func(*args, **kwargs) [ 636.034306] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 636.034306] env[62460]: raise e [ 636.034306] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 636.034306] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 636.034306] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 636.034306] env[62460]: created_port_ids = self._update_ports_for_instance( [ 636.034306] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 636.034306] env[62460]: with excutils.save_and_reraise_exception(): [ 636.034306] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 636.034306] env[62460]: self.force_reraise() [ 636.034306] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 636.034306] env[62460]: raise self.value [ 636.034306] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 636.034306] env[62460]: updated_port = self._update_port( [ 636.034306] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 636.034306] env[62460]: _ensure_no_port_binding_failure(port) [ 636.034306] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 636.034306] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 636.035299] env[62460]: nova.exception.PortBindingFailed: Binding failed for port 3221cf96-ce6e-40d8-bb3b-7ed671e0b4d1, please check neutron logs for more information. [ 636.035299] env[62460]: Removing descriptor: 19 [ 636.035299] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c33227d8-4250-4b3d-8e01-857e102558ea {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.044714] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1ffc068-c2f3-41fb-86c4-edf6145fbf05 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.068748] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3426ace3-0cff-4119-85fe-31e681a8597b could not be found. [ 636.069017] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 636.069372] env[62460]: INFO nova.compute.manager [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] Took 0.04 seconds to destroy the instance on the hypervisor. [ 636.069655] env[62460]: DEBUG oslo.service.loopingcall [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 636.069906] env[62460]: DEBUG nova.compute.manager [-] [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 636.070012] env[62460]: DEBUG nova.network.neutron [-] [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 636.095067] env[62460]: DEBUG nova.network.neutron [-] [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 636.311028] env[62460]: DEBUG nova.compute.manager [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 636.335810] env[62460]: DEBUG nova.virt.hardware [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T09:00:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='859ab47b-7eb3-44d8-bf9f-8e92ce0b7814',id=37,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1141297010',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 636.336069] env[62460]: DEBUG nova.virt.hardware [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 636.336233] env[62460]: DEBUG nova.virt.hardware [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 636.336925] env[62460]: DEBUG nova.virt.hardware [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 636.336925] env[62460]: DEBUG nova.virt.hardware [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 636.336925] env[62460]: DEBUG nova.virt.hardware [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 636.336925] env[62460]: DEBUG nova.virt.hardware [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 636.337166] env[62460]: DEBUG nova.virt.hardware [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 636.337243] env[62460]: DEBUG nova.virt.hardware [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 636.337408] env[62460]: DEBUG nova.virt.hardware [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 636.337579] env[62460]: DEBUG nova.virt.hardware [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 636.338442] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ae9a107-8d0a-4ea7-b557-defff0b181fa {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.341514] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 8c64f76d-cb18-41c5-9afa-e88af038f2b0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 636.348171] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11411726-b951-43c8-88b7-e50f3e85e9a8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.362316] env[62460]: ERROR nova.compute.manager [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3221cf96-ce6e-40d8-bb3b-7ed671e0b4d1, please check neutron logs for more information. [ 636.362316] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] Traceback (most recent call last): [ 636.362316] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 636.362316] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] yield resources [ 636.362316] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 636.362316] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] self.driver.spawn(context, instance, image_meta, [ 636.362316] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 636.362316] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 636.362316] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 636.362316] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] vm_ref = self.build_virtual_machine(instance, [ 636.362316] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 636.362692] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] vif_infos = vmwarevif.get_vif_info(self._session, [ 636.362692] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 636.362692] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] for vif in network_info: [ 636.362692] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 636.362692] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] return self._sync_wrapper(fn, *args, **kwargs) [ 636.362692] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 636.362692] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] self.wait() [ 636.362692] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 636.362692] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] self[:] = self._gt.wait() [ 636.362692] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 636.362692] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] return self._exit_event.wait() [ 636.362692] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 636.362692] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] current.throw(*self._exc) [ 636.363068] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 636.363068] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] result = function(*args, **kwargs) [ 636.363068] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 636.363068] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] return func(*args, **kwargs) [ 636.363068] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 636.363068] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] raise e [ 636.363068] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 636.363068] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] nwinfo = self.network_api.allocate_for_instance( [ 636.363068] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 636.363068] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] created_port_ids = self._update_ports_for_instance( [ 636.363068] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 636.363068] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] with excutils.save_and_reraise_exception(): [ 636.363068] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 636.363397] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] self.force_reraise() [ 636.363397] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 636.363397] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] raise self.value [ 636.363397] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 636.363397] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] updated_port = self._update_port( [ 636.363397] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 636.363397] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] _ensure_no_port_binding_failure(port) [ 636.363397] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 636.363397] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] raise exception.PortBindingFailed(port_id=port['id']) [ 636.363397] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] nova.exception.PortBindingFailed: Binding failed for port 3221cf96-ce6e-40d8-bb3b-7ed671e0b4d1, please check neutron logs for more information. [ 636.363397] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] [ 636.363397] env[62460]: INFO nova.compute.manager [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] Terminating instance [ 636.364494] env[62460]: DEBUG oslo_concurrency.lockutils [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Acquiring lock "refresh_cache-dba18c87-50a2-4ac8-8bd2-77f8f62b723c" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 636.364656] env[62460]: DEBUG oslo_concurrency.lockutils [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Acquired lock "refresh_cache-dba18c87-50a2-4ac8-8bd2-77f8f62b723c" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.364822] env[62460]: DEBUG nova.network.neutron [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 636.598429] env[62460]: DEBUG nova.network.neutron [-] [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.844464] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance a24776d9-9950-4c83-9641-9675cabd5fd3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 636.883940] env[62460]: DEBUG nova.network.neutron [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 636.954134] env[62460]: DEBUG nova.network.neutron [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.101092] env[62460]: INFO nova.compute.manager [-] [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] Took 1.03 seconds to deallocate network for instance. [ 637.103355] env[62460]: DEBUG nova.compute.claims [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 637.103578] env[62460]: DEBUG oslo_concurrency.lockutils [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 637.347413] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 61426715-7a38-475d-895a-0eb6d6040c66 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 637.456873] env[62460]: DEBUG oslo_concurrency.lockutils [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Releasing lock "refresh_cache-dba18c87-50a2-4ac8-8bd2-77f8f62b723c" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 637.457361] env[62460]: DEBUG nova.compute.manager [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 637.457602] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 637.457854] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1da5150c-0ca1-49ee-a429-03e320b6132b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.466916] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46a0f441-fccb-4f8d-802c-b5aeb0a4b3e8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.490021] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance dba18c87-50a2-4ac8-8bd2-77f8f62b723c could not be found. [ 637.490021] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 637.490021] env[62460]: INFO nova.compute.manager [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] Took 0.03 seconds to destroy the instance on the hypervisor. [ 637.490186] env[62460]: DEBUG oslo.service.loopingcall [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 637.490687] env[62460]: DEBUG nova.compute.manager [-] [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 637.490786] env[62460]: DEBUG nova.network.neutron [-] [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 637.506698] env[62460]: DEBUG nova.network.neutron [-] [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 637.850947] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance b6efad50-aa2e-49f9-9ce7-5fead31db7a8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 637.882403] env[62460]: DEBUG nova.compute.manager [req-c26f35a7-b68d-41ad-a579-ced18d7bc717 req-1d75e094-7ed0-46f2-bdd3-6110e5b32f8f service nova] [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] Received event network-changed-3221cf96-ce6e-40d8-bb3b-7ed671e0b4d1 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 637.882705] env[62460]: DEBUG nova.compute.manager [req-c26f35a7-b68d-41ad-a579-ced18d7bc717 req-1d75e094-7ed0-46f2-bdd3-6110e5b32f8f service nova] [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] Refreshing instance network info cache due to event network-changed-3221cf96-ce6e-40d8-bb3b-7ed671e0b4d1. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 637.883197] env[62460]: DEBUG oslo_concurrency.lockutils [req-c26f35a7-b68d-41ad-a579-ced18d7bc717 req-1d75e094-7ed0-46f2-bdd3-6110e5b32f8f service nova] Acquiring lock "refresh_cache-dba18c87-50a2-4ac8-8bd2-77f8f62b723c" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 637.883450] env[62460]: DEBUG oslo_concurrency.lockutils [req-c26f35a7-b68d-41ad-a579-ced18d7bc717 req-1d75e094-7ed0-46f2-bdd3-6110e5b32f8f service nova] Acquired lock "refresh_cache-dba18c87-50a2-4ac8-8bd2-77f8f62b723c" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 637.883720] env[62460]: DEBUG nova.network.neutron [req-c26f35a7-b68d-41ad-a579-ced18d7bc717 req-1d75e094-7ed0-46f2-bdd3-6110e5b32f8f service nova] [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] Refreshing network info cache for port 3221cf96-ce6e-40d8-bb3b-7ed671e0b4d1 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 638.009380] env[62460]: DEBUG nova.network.neutron [-] [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.354046] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 252df6ad-e29d-4596-bc99-4aae144bbcc5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 638.408493] env[62460]: DEBUG nova.network.neutron [req-c26f35a7-b68d-41ad-a579-ced18d7bc717 req-1d75e094-7ed0-46f2-bdd3-6110e5b32f8f service nova] [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 638.464757] env[62460]: DEBUG nova.network.neutron [req-c26f35a7-b68d-41ad-a579-ced18d7bc717 req-1d75e094-7ed0-46f2-bdd3-6110e5b32f8f service nova] [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.513236] env[62460]: INFO nova.compute.manager [-] [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] Took 1.02 seconds to deallocate network for instance. [ 638.515515] env[62460]: DEBUG nova.compute.claims [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 638.515698] env[62460]: DEBUG oslo_concurrency.lockutils [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 638.857545] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 02cb219f-989f-45b5-a9ba-d0b5a521cfa2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 638.967912] env[62460]: DEBUG oslo_concurrency.lockutils [req-c26f35a7-b68d-41ad-a579-ced18d7bc717 req-1d75e094-7ed0-46f2-bdd3-6110e5b32f8f service nova] Releasing lock "refresh_cache-dba18c87-50a2-4ac8-8bd2-77f8f62b723c" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 638.968199] env[62460]: DEBUG nova.compute.manager [req-c26f35a7-b68d-41ad-a579-ced18d7bc717 req-1d75e094-7ed0-46f2-bdd3-6110e5b32f8f service nova] [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] Received event network-vif-deleted-3221cf96-ce6e-40d8-bb3b-7ed671e0b4d1 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 639.361800] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 578627be-e695-4953-8d0e-9763d12b9a28 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 639.864853] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 724a8083-2bde-483e-bd5a-a928def284bf has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 640.367848] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 0d29a4ce-6689-4d0c-8144-e05853eb3a60 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 640.870947] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 02b7a3c3-bdfd-47d9-acf7-afd7725cea64 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 641.373949] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 641.879297] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 770b6195-ab31-4df0-84b0-e8382732ec32 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 642.382798] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 642.886079] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance c37f1b55-7a06-4c93-9492-d2bd6d2bfe83 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 643.389428] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance ec2297bb-1d29-4ea9-90f7-f4f39716c103 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 643.894194] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 0e00ca48-f9e1-4999-aad5-d7965f7ddc28 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 644.396734] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 644.899701] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance c4c5abc4-a5ed-4337-a930-d37d579819f8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 645.403172] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 4ed90d16-81a6-4dbd-8936-0e137151171f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 645.906226] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 5214f4a1-3e28-41bf-88d9-161511385e1b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 645.906508] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=62460) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 645.906638] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1472MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=5 pci_stats=[] {{(pid=62460) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 646.186260] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c79c4fc4-dc94-40f5-baad-e22ca4d3aca0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.193465] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed70ee98-deff-44b7-b5cc-3d415fd02bd7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.222641] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad5a9623-c4dc-422e-9131-6241e3202282 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.229192] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e20329a-4d5c-4ef1-bd4f-345bf27b18b1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.241749] env[62460]: DEBUG nova.compute.provider_tree [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 646.744663] env[62460]: DEBUG nova.scheduler.client.report [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 647.250065] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62460) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 647.250065] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 12.961s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 647.250416] env[62460]: DEBUG oslo_concurrency.lockutils [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.943s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 647.251740] env[62460]: INFO nova.compute.claims [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 648.537055] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7778bc6-735f-4ed0-81b4-73b6d93c36b6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.544123] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5008c545-8842-4df9-bd54-e4ba4ae5283b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.574896] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70225b20-36fc-4a5f-9f92-c497fbde7ce9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.582558] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13a2c05d-6cd5-49b9-bc44-d863d45a8617 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.595985] env[62460]: DEBUG nova.compute.provider_tree [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 649.103376] env[62460]: DEBUG nova.scheduler.client.report [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 649.607964] env[62460]: DEBUG oslo_concurrency.lockutils [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.358s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 649.608558] env[62460]: DEBUG nova.compute.manager [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 649.611926] env[62460]: DEBUG oslo_concurrency.lockutils [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.404s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 649.613588] env[62460]: INFO nova.compute.claims [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 650.113373] env[62460]: DEBUG nova.compute.utils [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 650.114815] env[62460]: DEBUG nova.compute.manager [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 650.114990] env[62460]: DEBUG nova.network.neutron [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 650.156416] env[62460]: DEBUG nova.policy [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8e09d51001bc4e40b4d35db1511d1a6e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '210ef0c43440478b9dcc5360edf52241', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 650.429783] env[62460]: DEBUG nova.network.neutron [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] Successfully created port: 759c443f-21d6-4fef-a27b-6fe32cc20a63 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 650.617767] env[62460]: DEBUG nova.compute.manager [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 650.954923] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d933bb54-39e7-4944-8c75-d4cfaa4340b7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.963173] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39f85750-2e71-47b6-94cb-f241030efd66 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.995018] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4a5ac85-a09b-4ad1-88a7-75cce96cf248 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.001040] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bcdf990-222a-4cea-bb85-7e5935739994 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.014333] env[62460]: DEBUG nova.compute.provider_tree [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 651.236801] env[62460]: DEBUG nova.compute.manager [req-2c0ddcf8-130a-4ec5-9304-2bdbd9301c39 req-8fe40e83-8806-4129-a873-5a3d1f9cdbc7 service nova] [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] Received event network-changed-759c443f-21d6-4fef-a27b-6fe32cc20a63 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 651.237024] env[62460]: DEBUG nova.compute.manager [req-2c0ddcf8-130a-4ec5-9304-2bdbd9301c39 req-8fe40e83-8806-4129-a873-5a3d1f9cdbc7 service nova] [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] Refreshing instance network info cache due to event network-changed-759c443f-21d6-4fef-a27b-6fe32cc20a63. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 651.237939] env[62460]: DEBUG oslo_concurrency.lockutils [req-2c0ddcf8-130a-4ec5-9304-2bdbd9301c39 req-8fe40e83-8806-4129-a873-5a3d1f9cdbc7 service nova] Acquiring lock "refresh_cache-01a3e1be-d29b-45d7-987e-66a4395ae2a4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 651.237939] env[62460]: DEBUG oslo_concurrency.lockutils [req-2c0ddcf8-130a-4ec5-9304-2bdbd9301c39 req-8fe40e83-8806-4129-a873-5a3d1f9cdbc7 service nova] Acquired lock "refresh_cache-01a3e1be-d29b-45d7-987e-66a4395ae2a4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 651.237939] env[62460]: DEBUG nova.network.neutron [req-2c0ddcf8-130a-4ec5-9304-2bdbd9301c39 req-8fe40e83-8806-4129-a873-5a3d1f9cdbc7 service nova] [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] Refreshing network info cache for port 759c443f-21d6-4fef-a27b-6fe32cc20a63 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 651.418969] env[62460]: ERROR nova.compute.manager [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 759c443f-21d6-4fef-a27b-6fe32cc20a63, please check neutron logs for more information. [ 651.418969] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 651.418969] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 651.418969] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 651.418969] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 651.418969] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 651.418969] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 651.418969] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 651.418969] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 651.418969] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 651.418969] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 651.418969] env[62460]: ERROR nova.compute.manager raise self.value [ 651.418969] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 651.418969] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 651.418969] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 651.418969] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 651.419455] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 651.419455] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 651.419455] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 759c443f-21d6-4fef-a27b-6fe32cc20a63, please check neutron logs for more information. [ 651.419455] env[62460]: ERROR nova.compute.manager [ 651.419455] env[62460]: Traceback (most recent call last): [ 651.419455] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 651.419455] env[62460]: listener.cb(fileno) [ 651.419455] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 651.419455] env[62460]: result = function(*args, **kwargs) [ 651.419455] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 651.419455] env[62460]: return func(*args, **kwargs) [ 651.419455] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 651.419455] env[62460]: raise e [ 651.419455] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 651.419455] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 651.419455] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 651.419455] env[62460]: created_port_ids = self._update_ports_for_instance( [ 651.419455] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 651.419455] env[62460]: with excutils.save_and_reraise_exception(): [ 651.419455] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 651.419455] env[62460]: self.force_reraise() [ 651.419455] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 651.419455] env[62460]: raise self.value [ 651.419455] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 651.419455] env[62460]: updated_port = self._update_port( [ 651.419455] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 651.419455] env[62460]: _ensure_no_port_binding_failure(port) [ 651.419455] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 651.419455] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 651.420222] env[62460]: nova.exception.PortBindingFailed: Binding failed for port 759c443f-21d6-4fef-a27b-6fe32cc20a63, please check neutron logs for more information. [ 651.420222] env[62460]: Removing descriptor: 19 [ 651.517418] env[62460]: DEBUG nova.scheduler.client.report [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 651.629368] env[62460]: DEBUG nova.compute.manager [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 651.655112] env[62460]: DEBUG nova.virt.hardware [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 651.655361] env[62460]: DEBUG nova.virt.hardware [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 651.655522] env[62460]: DEBUG nova.virt.hardware [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 651.655711] env[62460]: DEBUG nova.virt.hardware [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 651.655858] env[62460]: DEBUG nova.virt.hardware [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 651.656016] env[62460]: DEBUG nova.virt.hardware [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 651.656413] env[62460]: DEBUG nova.virt.hardware [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 651.656413] env[62460]: DEBUG nova.virt.hardware [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 651.656559] env[62460]: DEBUG nova.virt.hardware [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 651.656725] env[62460]: DEBUG nova.virt.hardware [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 651.656897] env[62460]: DEBUG nova.virt.hardware [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 651.658745] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b86a8e8-7a54-4a7e-aebc-e50fe221d041 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.667251] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bde3aa07-d8f4-401c-9f57-b4dfcf9aadd5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.681022] env[62460]: ERROR nova.compute.manager [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 759c443f-21d6-4fef-a27b-6fe32cc20a63, please check neutron logs for more information. [ 651.681022] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] Traceback (most recent call last): [ 651.681022] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 651.681022] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] yield resources [ 651.681022] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 651.681022] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] self.driver.spawn(context, instance, image_meta, [ 651.681022] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 651.681022] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 651.681022] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 651.681022] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] vm_ref = self.build_virtual_machine(instance, [ 651.681022] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 651.681362] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] vif_infos = vmwarevif.get_vif_info(self._session, [ 651.681362] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 651.681362] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] for vif in network_info: [ 651.681362] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 651.681362] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] return self._sync_wrapper(fn, *args, **kwargs) [ 651.681362] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 651.681362] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] self.wait() [ 651.681362] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 651.681362] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] self[:] = self._gt.wait() [ 651.681362] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 651.681362] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] return self._exit_event.wait() [ 651.681362] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 651.681362] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] current.throw(*self._exc) [ 651.681744] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 651.681744] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] result = function(*args, **kwargs) [ 651.681744] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 651.681744] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] return func(*args, **kwargs) [ 651.681744] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 651.681744] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] raise e [ 651.681744] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 651.681744] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] nwinfo = self.network_api.allocate_for_instance( [ 651.681744] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 651.681744] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] created_port_ids = self._update_ports_for_instance( [ 651.681744] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 651.681744] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] with excutils.save_and_reraise_exception(): [ 651.681744] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 651.682125] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] self.force_reraise() [ 651.682125] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 651.682125] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] raise self.value [ 651.682125] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 651.682125] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] updated_port = self._update_port( [ 651.682125] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 651.682125] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] _ensure_no_port_binding_failure(port) [ 651.682125] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 651.682125] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] raise exception.PortBindingFailed(port_id=port['id']) [ 651.682125] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] nova.exception.PortBindingFailed: Binding failed for port 759c443f-21d6-4fef-a27b-6fe32cc20a63, please check neutron logs for more information. [ 651.682125] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] [ 651.682125] env[62460]: INFO nova.compute.manager [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] Terminating instance [ 651.686248] env[62460]: DEBUG oslo_concurrency.lockutils [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Acquiring lock "refresh_cache-01a3e1be-d29b-45d7-987e-66a4395ae2a4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 651.756063] env[62460]: DEBUG nova.network.neutron [req-2c0ddcf8-130a-4ec5-9304-2bdbd9301c39 req-8fe40e83-8806-4129-a873-5a3d1f9cdbc7 service nova] [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 651.847729] env[62460]: DEBUG nova.network.neutron [req-2c0ddcf8-130a-4ec5-9304-2bdbd9301c39 req-8fe40e83-8806-4129-a873-5a3d1f9cdbc7 service nova] [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 652.024052] env[62460]: DEBUG oslo_concurrency.lockutils [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.412s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 652.024052] env[62460]: DEBUG nova.compute.manager [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 652.026634] env[62460]: DEBUG oslo_concurrency.lockutils [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.353s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 652.028050] env[62460]: INFO nova.compute.claims [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 652.350643] env[62460]: DEBUG oslo_concurrency.lockutils [req-2c0ddcf8-130a-4ec5-9304-2bdbd9301c39 req-8fe40e83-8806-4129-a873-5a3d1f9cdbc7 service nova] Releasing lock "refresh_cache-01a3e1be-d29b-45d7-987e-66a4395ae2a4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 652.350951] env[62460]: DEBUG oslo_concurrency.lockutils [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Acquired lock "refresh_cache-01a3e1be-d29b-45d7-987e-66a4395ae2a4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 652.351156] env[62460]: DEBUG nova.network.neutron [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 652.532514] env[62460]: DEBUG nova.compute.utils [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 652.535679] env[62460]: DEBUG nova.compute.manager [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 652.535850] env[62460]: DEBUG nova.network.neutron [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 652.576500] env[62460]: DEBUG nova.policy [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0f21627a28854b368512373c5931e51f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b76c231415f742e0a7a05dd04eee5fbc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 652.867169] env[62460]: DEBUG nova.network.neutron [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] Successfully created port: 4fa12c2e-8d85-4ab5-bec3-f905bdec71d5 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 652.900460] env[62460]: DEBUG nova.network.neutron [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 652.983095] env[62460]: DEBUG nova.network.neutron [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.036819] env[62460]: DEBUG nova.compute.manager [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 653.264242] env[62460]: DEBUG nova.compute.manager [req-df751554-b6dc-4fa3-862b-c18bb68b93ad req-7735533b-68ce-4d29-b9a6-46c081aa21e1 service nova] [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] Received event network-vif-deleted-759c443f-21d6-4fef-a27b-6fe32cc20a63 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 653.372808] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c97d9580-d046-4423-8e13-c5c66788fa50 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.380454] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59acdc79-1d14-4fdf-8303-ed586e758920 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.415301] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebb360b2-c72f-4b1e-a47c-6e57694f45ee {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.424982] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-461d414c-0d2e-4d1f-b5d1-2dc1d8b1d176 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.438556] env[62460]: DEBUG nova.compute.provider_tree [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 653.492625] env[62460]: DEBUG oslo_concurrency.lockutils [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Releasing lock "refresh_cache-01a3e1be-d29b-45d7-987e-66a4395ae2a4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 653.492625] env[62460]: DEBUG nova.compute.manager [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 653.492625] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 653.492625] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dd99a64d-5fc4-40f1-bbdc-912569980ffc {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.500937] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-359c6e9b-82c3-42ca-a8fc-d66596ccaf51 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.522543] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 01a3e1be-d29b-45d7-987e-66a4395ae2a4 could not be found. [ 653.522543] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 653.523131] env[62460]: INFO nova.compute.manager [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] Took 0.03 seconds to destroy the instance on the hypervisor. [ 653.523131] env[62460]: DEBUG oslo.service.loopingcall [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 653.523301] env[62460]: DEBUG nova.compute.manager [-] [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 653.523396] env[62460]: DEBUG nova.network.neutron [-] [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 653.550020] env[62460]: DEBUG nova.network.neutron [-] [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 653.826159] env[62460]: ERROR nova.compute.manager [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4fa12c2e-8d85-4ab5-bec3-f905bdec71d5, please check neutron logs for more information. [ 653.826159] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 653.826159] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 653.826159] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 653.826159] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 653.826159] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 653.826159] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 653.826159] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 653.826159] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 653.826159] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 653.826159] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 653.826159] env[62460]: ERROR nova.compute.manager raise self.value [ 653.826159] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 653.826159] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 653.826159] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 653.826159] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 653.826698] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 653.826698] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 653.826698] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4fa12c2e-8d85-4ab5-bec3-f905bdec71d5, please check neutron logs for more information. [ 653.826698] env[62460]: ERROR nova.compute.manager [ 653.826698] env[62460]: Traceback (most recent call last): [ 653.826698] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 653.826698] env[62460]: listener.cb(fileno) [ 653.826698] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 653.826698] env[62460]: result = function(*args, **kwargs) [ 653.826698] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 653.826698] env[62460]: return func(*args, **kwargs) [ 653.826698] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 653.826698] env[62460]: raise e [ 653.826698] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 653.826698] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 653.826698] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 653.826698] env[62460]: created_port_ids = self._update_ports_for_instance( [ 653.826698] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 653.826698] env[62460]: with excutils.save_and_reraise_exception(): [ 653.826698] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 653.826698] env[62460]: self.force_reraise() [ 653.826698] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 653.826698] env[62460]: raise self.value [ 653.826698] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 653.826698] env[62460]: updated_port = self._update_port( [ 653.826698] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 653.826698] env[62460]: _ensure_no_port_binding_failure(port) [ 653.826698] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 653.826698] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 653.827874] env[62460]: nova.exception.PortBindingFailed: Binding failed for port 4fa12c2e-8d85-4ab5-bec3-f905bdec71d5, please check neutron logs for more information. [ 653.827874] env[62460]: Removing descriptor: 19 [ 653.942127] env[62460]: DEBUG nova.scheduler.client.report [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 654.047644] env[62460]: DEBUG nova.compute.manager [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 654.051300] env[62460]: DEBUG nova.network.neutron [-] [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 654.074952] env[62460]: DEBUG nova.virt.hardware [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 654.075264] env[62460]: DEBUG nova.virt.hardware [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 654.075458] env[62460]: DEBUG nova.virt.hardware [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 654.075781] env[62460]: DEBUG nova.virt.hardware [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 654.075963] env[62460]: DEBUG nova.virt.hardware [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 654.076088] env[62460]: DEBUG nova.virt.hardware [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 654.076346] env[62460]: DEBUG nova.virt.hardware [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 654.076547] env[62460]: DEBUG nova.virt.hardware [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 654.076782] env[62460]: DEBUG nova.virt.hardware [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 654.076956] env[62460]: DEBUG nova.virt.hardware [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 654.077199] env[62460]: DEBUG nova.virt.hardware [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 654.078433] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a823b98-fd5f-4825-b328-ee5c01217261 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.087408] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccf8b5f1-3b43-4767-97b2-31493c8421ea {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.101042] env[62460]: ERROR nova.compute.manager [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4fa12c2e-8d85-4ab5-bec3-f905bdec71d5, please check neutron logs for more information. [ 654.101042] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] Traceback (most recent call last): [ 654.101042] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 654.101042] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] yield resources [ 654.101042] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 654.101042] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] self.driver.spawn(context, instance, image_meta, [ 654.101042] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 654.101042] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 654.101042] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 654.101042] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] vm_ref = self.build_virtual_machine(instance, [ 654.101042] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 654.101446] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] vif_infos = vmwarevif.get_vif_info(self._session, [ 654.101446] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 654.101446] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] for vif in network_info: [ 654.101446] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 654.101446] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] return self._sync_wrapper(fn, *args, **kwargs) [ 654.101446] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 654.101446] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] self.wait() [ 654.101446] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 654.101446] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] self[:] = self._gt.wait() [ 654.101446] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 654.101446] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] return self._exit_event.wait() [ 654.101446] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 654.101446] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] current.throw(*self._exc) [ 654.101832] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 654.101832] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] result = function(*args, **kwargs) [ 654.101832] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 654.101832] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] return func(*args, **kwargs) [ 654.101832] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 654.101832] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] raise e [ 654.101832] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 654.101832] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] nwinfo = self.network_api.allocate_for_instance( [ 654.101832] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 654.101832] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] created_port_ids = self._update_ports_for_instance( [ 654.101832] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 654.101832] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] with excutils.save_and_reraise_exception(): [ 654.101832] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 654.102188] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] self.force_reraise() [ 654.102188] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 654.102188] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] raise self.value [ 654.102188] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 654.102188] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] updated_port = self._update_port( [ 654.102188] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 654.102188] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] _ensure_no_port_binding_failure(port) [ 654.102188] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 654.102188] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] raise exception.PortBindingFailed(port_id=port['id']) [ 654.102188] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] nova.exception.PortBindingFailed: Binding failed for port 4fa12c2e-8d85-4ab5-bec3-f905bdec71d5, please check neutron logs for more information. [ 654.102188] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] [ 654.102188] env[62460]: INFO nova.compute.manager [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] Terminating instance [ 654.103367] env[62460]: DEBUG oslo_concurrency.lockutils [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Acquiring lock "refresh_cache-8c64f76d-cb18-41c5-9afa-e88af038f2b0" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 654.103530] env[62460]: DEBUG oslo_concurrency.lockutils [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Acquired lock "refresh_cache-8c64f76d-cb18-41c5-9afa-e88af038f2b0" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 654.103701] env[62460]: DEBUG nova.network.neutron [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 654.448888] env[62460]: DEBUG oslo_concurrency.lockutils [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.422s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 654.450122] env[62460]: DEBUG nova.compute.manager [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 654.452894] env[62460]: DEBUG oslo_concurrency.lockutils [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.047s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 654.455021] env[62460]: INFO nova.compute.claims [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] [instance: 61426715-7a38-475d-895a-0eb6d6040c66] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 654.553920] env[62460]: INFO nova.compute.manager [-] [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] Took 1.03 seconds to deallocate network for instance. [ 654.556306] env[62460]: DEBUG nova.compute.claims [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 654.556488] env[62460]: DEBUG oslo_concurrency.lockutils [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 654.624460] env[62460]: DEBUG nova.network.neutron [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 654.698694] env[62460]: DEBUG nova.network.neutron [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 654.962145] env[62460]: DEBUG nova.compute.utils [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 654.963567] env[62460]: DEBUG nova.compute.manager [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 654.963756] env[62460]: DEBUG nova.network.neutron [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 655.015280] env[62460]: DEBUG nova.policy [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd7ac705ce4dc4f42b89cf70ac3316f71', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '335cfe76a8c44686891e17906c96d158', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 655.201724] env[62460]: DEBUG oslo_concurrency.lockutils [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Releasing lock "refresh_cache-8c64f76d-cb18-41c5-9afa-e88af038f2b0" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 655.202237] env[62460]: DEBUG nova.compute.manager [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 655.202438] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 655.202751] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-45852281-1bbd-474e-910b-ee4d6c77f493 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.217609] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecc9b1d4-cda5-4d23-80e5-0925ee57f48c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.238189] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8c64f76d-cb18-41c5-9afa-e88af038f2b0 could not be found. [ 655.238555] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 655.238857] env[62460]: INFO nova.compute.manager [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] Took 0.04 seconds to destroy the instance on the hypervisor. [ 655.239212] env[62460]: DEBUG oslo.service.loopingcall [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 655.239550] env[62460]: DEBUG nova.compute.manager [-] [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 655.239753] env[62460]: DEBUG nova.network.neutron [-] [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 655.265295] env[62460]: DEBUG nova.network.neutron [-] [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 655.313920] env[62460]: DEBUG nova.compute.manager [req-c46182ed-1ba2-4c22-9710-5d61e924f095 req-6ee1e9c9-9272-42e6-a087-8bf483cbdfe7 service nova] [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] Received event network-changed-4fa12c2e-8d85-4ab5-bec3-f905bdec71d5 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 655.314165] env[62460]: DEBUG nova.compute.manager [req-c46182ed-1ba2-4c22-9710-5d61e924f095 req-6ee1e9c9-9272-42e6-a087-8bf483cbdfe7 service nova] [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] Refreshing instance network info cache due to event network-changed-4fa12c2e-8d85-4ab5-bec3-f905bdec71d5. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 655.314349] env[62460]: DEBUG oslo_concurrency.lockutils [req-c46182ed-1ba2-4c22-9710-5d61e924f095 req-6ee1e9c9-9272-42e6-a087-8bf483cbdfe7 service nova] Acquiring lock "refresh_cache-8c64f76d-cb18-41c5-9afa-e88af038f2b0" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 655.314519] env[62460]: DEBUG oslo_concurrency.lockutils [req-c46182ed-1ba2-4c22-9710-5d61e924f095 req-6ee1e9c9-9272-42e6-a087-8bf483cbdfe7 service nova] Acquired lock "refresh_cache-8c64f76d-cb18-41c5-9afa-e88af038f2b0" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 655.314639] env[62460]: DEBUG nova.network.neutron [req-c46182ed-1ba2-4c22-9710-5d61e924f095 req-6ee1e9c9-9272-42e6-a087-8bf483cbdfe7 service nova] [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] Refreshing network info cache for port 4fa12c2e-8d85-4ab5-bec3-f905bdec71d5 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 655.368087] env[62460]: DEBUG nova.network.neutron [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] Successfully created port: 07d0ef21-dd97-42e1-8b07-67b1feade067 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 655.467741] env[62460]: DEBUG nova.compute.manager [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 655.770260] env[62460]: DEBUG nova.network.neutron [-] [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 655.826931] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12b7189d-1bad-477a-8f3c-6a91c0fad82b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.833135] env[62460]: DEBUG nova.network.neutron [req-c46182ed-1ba2-4c22-9710-5d61e924f095 req-6ee1e9c9-9272-42e6-a087-8bf483cbdfe7 service nova] [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 655.837305] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6589d6b-4433-4749-a6e0-b0e446fcc166 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.866473] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10a3244d-74df-4437-96d4-2560dc33343b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.873798] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99ae7464-5d5f-451b-bb69-0f5d5d85a368 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.891025] env[62460]: DEBUG nova.compute.provider_tree [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 655.926455] env[62460]: DEBUG nova.network.neutron [req-c46182ed-1ba2-4c22-9710-5d61e924f095 req-6ee1e9c9-9272-42e6-a087-8bf483cbdfe7 service nova] [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.271732] env[62460]: INFO nova.compute.manager [-] [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] Took 1.03 seconds to deallocate network for instance. [ 656.274121] env[62460]: DEBUG nova.compute.claims [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 656.274300] env[62460]: DEBUG oslo_concurrency.lockutils [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 656.339754] env[62460]: ERROR nova.compute.manager [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 07d0ef21-dd97-42e1-8b07-67b1feade067, please check neutron logs for more information. [ 656.339754] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 656.339754] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 656.339754] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 656.339754] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 656.339754] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 656.339754] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 656.339754] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 656.339754] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 656.339754] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 656.339754] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 656.339754] env[62460]: ERROR nova.compute.manager raise self.value [ 656.339754] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 656.339754] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 656.339754] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 656.339754] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 656.340268] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 656.340268] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 656.340268] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 07d0ef21-dd97-42e1-8b07-67b1feade067, please check neutron logs for more information. [ 656.340268] env[62460]: ERROR nova.compute.manager [ 656.340268] env[62460]: Traceback (most recent call last): [ 656.340268] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 656.340268] env[62460]: listener.cb(fileno) [ 656.340268] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 656.340268] env[62460]: result = function(*args, **kwargs) [ 656.340268] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 656.340268] env[62460]: return func(*args, **kwargs) [ 656.340268] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 656.340268] env[62460]: raise e [ 656.340268] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 656.340268] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 656.340268] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 656.340268] env[62460]: created_port_ids = self._update_ports_for_instance( [ 656.340268] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 656.340268] env[62460]: with excutils.save_and_reraise_exception(): [ 656.340268] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 656.340268] env[62460]: self.force_reraise() [ 656.340268] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 656.340268] env[62460]: raise self.value [ 656.340268] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 656.340268] env[62460]: updated_port = self._update_port( [ 656.340268] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 656.340268] env[62460]: _ensure_no_port_binding_failure(port) [ 656.340268] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 656.340268] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 656.341100] env[62460]: nova.exception.PortBindingFailed: Binding failed for port 07d0ef21-dd97-42e1-8b07-67b1feade067, please check neutron logs for more information. [ 656.341100] env[62460]: Removing descriptor: 19 [ 656.391731] env[62460]: DEBUG nova.scheduler.client.report [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 656.429043] env[62460]: DEBUG oslo_concurrency.lockutils [req-c46182ed-1ba2-4c22-9710-5d61e924f095 req-6ee1e9c9-9272-42e6-a087-8bf483cbdfe7 service nova] Releasing lock "refresh_cache-8c64f76d-cb18-41c5-9afa-e88af038f2b0" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 656.429330] env[62460]: DEBUG nova.compute.manager [req-c46182ed-1ba2-4c22-9710-5d61e924f095 req-6ee1e9c9-9272-42e6-a087-8bf483cbdfe7 service nova] [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] Received event network-vif-deleted-4fa12c2e-8d85-4ab5-bec3-f905bdec71d5 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 656.477154] env[62460]: DEBUG nova.compute.manager [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 656.505027] env[62460]: DEBUG nova.virt.hardware [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 656.505292] env[62460]: DEBUG nova.virt.hardware [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 656.505451] env[62460]: DEBUG nova.virt.hardware [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 656.505629] env[62460]: DEBUG nova.virt.hardware [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 656.505772] env[62460]: DEBUG nova.virt.hardware [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 656.505917] env[62460]: DEBUG nova.virt.hardware [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 656.506181] env[62460]: DEBUG nova.virt.hardware [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 656.506290] env[62460]: DEBUG nova.virt.hardware [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 656.506454] env[62460]: DEBUG nova.virt.hardware [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 656.506612] env[62460]: DEBUG nova.virt.hardware [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 656.506781] env[62460]: DEBUG nova.virt.hardware [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 656.507681] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a4e925b-f7af-4dda-b18b-61e8a0403b6c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.515565] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04d966a5-7ab2-4379-92af-73dfd429fb44 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.529202] env[62460]: ERROR nova.compute.manager [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 07d0ef21-dd97-42e1-8b07-67b1feade067, please check neutron logs for more information. [ 656.529202] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] Traceback (most recent call last): [ 656.529202] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 656.529202] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] yield resources [ 656.529202] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 656.529202] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] self.driver.spawn(context, instance, image_meta, [ 656.529202] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 656.529202] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 656.529202] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 656.529202] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] vm_ref = self.build_virtual_machine(instance, [ 656.529202] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 656.529556] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] vif_infos = vmwarevif.get_vif_info(self._session, [ 656.529556] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 656.529556] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] for vif in network_info: [ 656.529556] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 656.529556] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] return self._sync_wrapper(fn, *args, **kwargs) [ 656.529556] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 656.529556] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] self.wait() [ 656.529556] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 656.529556] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] self[:] = self._gt.wait() [ 656.529556] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 656.529556] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] return self._exit_event.wait() [ 656.529556] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 656.529556] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] current.throw(*self._exc) [ 656.529903] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 656.529903] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] result = function(*args, **kwargs) [ 656.529903] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 656.529903] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] return func(*args, **kwargs) [ 656.529903] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 656.529903] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] raise e [ 656.529903] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 656.529903] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] nwinfo = self.network_api.allocate_for_instance( [ 656.529903] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 656.529903] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] created_port_ids = self._update_ports_for_instance( [ 656.529903] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 656.529903] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] with excutils.save_and_reraise_exception(): [ 656.529903] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 656.530248] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] self.force_reraise() [ 656.530248] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 656.530248] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] raise self.value [ 656.530248] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 656.530248] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] updated_port = self._update_port( [ 656.530248] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 656.530248] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] _ensure_no_port_binding_failure(port) [ 656.530248] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 656.530248] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] raise exception.PortBindingFailed(port_id=port['id']) [ 656.530248] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] nova.exception.PortBindingFailed: Binding failed for port 07d0ef21-dd97-42e1-8b07-67b1feade067, please check neutron logs for more information. [ 656.530248] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] [ 656.530248] env[62460]: INFO nova.compute.manager [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] Terminating instance [ 656.531501] env[62460]: DEBUG oslo_concurrency.lockutils [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Acquiring lock "refresh_cache-a24776d9-9950-4c83-9641-9675cabd5fd3" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 656.531671] env[62460]: DEBUG oslo_concurrency.lockutils [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Acquired lock "refresh_cache-a24776d9-9950-4c83-9641-9675cabd5fd3" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 656.531842] env[62460]: DEBUG nova.network.neutron [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 656.897060] env[62460]: DEBUG oslo_concurrency.lockutils [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.444s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 656.897658] env[62460]: DEBUG nova.compute.manager [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] [instance: 61426715-7a38-475d-895a-0eb6d6040c66] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 656.900687] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 28.314s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 657.049499] env[62460]: DEBUG nova.network.neutron [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 657.121738] env[62460]: DEBUG nova.network.neutron [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 657.347451] env[62460]: DEBUG nova.compute.manager [req-f5007b40-e61f-4075-8dbc-6ebf01e5d84b req-3fdede88-bec9-417c-a680-9de54cd39d12 service nova] [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] Received event network-changed-07d0ef21-dd97-42e1-8b07-67b1feade067 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 657.347694] env[62460]: DEBUG nova.compute.manager [req-f5007b40-e61f-4075-8dbc-6ebf01e5d84b req-3fdede88-bec9-417c-a680-9de54cd39d12 service nova] [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] Refreshing instance network info cache due to event network-changed-07d0ef21-dd97-42e1-8b07-67b1feade067. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 657.348314] env[62460]: DEBUG oslo_concurrency.lockutils [req-f5007b40-e61f-4075-8dbc-6ebf01e5d84b req-3fdede88-bec9-417c-a680-9de54cd39d12 service nova] Acquiring lock "refresh_cache-a24776d9-9950-4c83-9641-9675cabd5fd3" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 657.407906] env[62460]: DEBUG nova.compute.utils [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 657.412243] env[62460]: DEBUG nova.compute.manager [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] [instance: 61426715-7a38-475d-895a-0eb6d6040c66] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 657.412419] env[62460]: DEBUG nova.network.neutron [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] [instance: 61426715-7a38-475d-895a-0eb6d6040c66] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 657.450434] env[62460]: DEBUG nova.policy [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4b66806dce724109b7965ff8972469f2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aaf588930aa94cefb68c1c52ca8487d9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 657.626710] env[62460]: DEBUG oslo_concurrency.lockutils [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Releasing lock "refresh_cache-a24776d9-9950-4c83-9641-9675cabd5fd3" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 657.626710] env[62460]: DEBUG nova.compute.manager [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 657.626710] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 657.626710] env[62460]: DEBUG oslo_concurrency.lockutils [req-f5007b40-e61f-4075-8dbc-6ebf01e5d84b req-3fdede88-bec9-417c-a680-9de54cd39d12 service nova] Acquired lock "refresh_cache-a24776d9-9950-4c83-9641-9675cabd5fd3" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 657.626710] env[62460]: DEBUG nova.network.neutron [req-f5007b40-e61f-4075-8dbc-6ebf01e5d84b req-3fdede88-bec9-417c-a680-9de54cd39d12 service nova] [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] Refreshing network info cache for port 07d0ef21-dd97-42e1-8b07-67b1feade067 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 657.627859] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2dfb4ebd-bd92-4491-9c0b-e7849e19e4b5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.644548] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c6211eb-48f0-49ba-af46-fac9a4b50bdd {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.671931] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a24776d9-9950-4c83-9641-9675cabd5fd3 could not be found. [ 657.671931] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 657.672150] env[62460]: INFO nova.compute.manager [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] Took 0.05 seconds to destroy the instance on the hypervisor. [ 657.672393] env[62460]: DEBUG oslo.service.loopingcall [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 657.674805] env[62460]: DEBUG nova.compute.manager [-] [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 657.674909] env[62460]: DEBUG nova.network.neutron [-] [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 657.705508] env[62460]: DEBUG nova.network.neutron [-] [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 657.759584] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d442c229-4e0f-42cb-a011-9bfafd5fd68b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.763986] env[62460]: DEBUG nova.network.neutron [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] [instance: 61426715-7a38-475d-895a-0eb6d6040c66] Successfully created port: 8ecec905-4891-4a60-a063-ef5f68a1373b {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 657.768553] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c130a7e-33cf-48c4-a4b5-99661169d4a1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.798542] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d887288e-0fac-4ff7-b9d2-baede1b43239 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.806604] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10979d71-8545-4586-915e-251ad5f05beb {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.820496] env[62460]: DEBUG nova.compute.provider_tree [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 657.913614] env[62460]: DEBUG nova.compute.manager [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] [instance: 61426715-7a38-475d-895a-0eb6d6040c66] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 658.147788] env[62460]: DEBUG nova.network.neutron [req-f5007b40-e61f-4075-8dbc-6ebf01e5d84b req-3fdede88-bec9-417c-a680-9de54cd39d12 service nova] [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 658.208161] env[62460]: DEBUG nova.network.neutron [-] [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.240455] env[62460]: DEBUG nova.network.neutron [req-f5007b40-e61f-4075-8dbc-6ebf01e5d84b req-3fdede88-bec9-417c-a680-9de54cd39d12 service nova] [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.323589] env[62460]: DEBUG nova.scheduler.client.report [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 658.714667] env[62460]: INFO nova.compute.manager [-] [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] Took 1.04 seconds to deallocate network for instance. [ 658.716852] env[62460]: DEBUG nova.compute.claims [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 658.716852] env[62460]: DEBUG oslo_concurrency.lockutils [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 658.731415] env[62460]: ERROR nova.compute.manager [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8ecec905-4891-4a60-a063-ef5f68a1373b, please check neutron logs for more information. [ 658.731415] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 658.731415] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 658.731415] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 658.731415] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 658.731415] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 658.731415] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 658.731415] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 658.731415] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 658.731415] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 658.731415] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 658.731415] env[62460]: ERROR nova.compute.manager raise self.value [ 658.731415] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 658.731415] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 658.731415] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 658.731415] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 658.731962] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 658.731962] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 658.731962] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8ecec905-4891-4a60-a063-ef5f68a1373b, please check neutron logs for more information. [ 658.731962] env[62460]: ERROR nova.compute.manager [ 658.732526] env[62460]: Traceback (most recent call last): [ 658.732526] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 658.732526] env[62460]: listener.cb(fileno) [ 658.732526] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 658.732526] env[62460]: result = function(*args, **kwargs) [ 658.732526] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 658.732526] env[62460]: return func(*args, **kwargs) [ 658.732526] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 658.732526] env[62460]: raise e [ 658.732526] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 658.732526] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 658.732526] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 658.732526] env[62460]: created_port_ids = self._update_ports_for_instance( [ 658.732526] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 658.732526] env[62460]: with excutils.save_and_reraise_exception(): [ 658.732526] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 658.732526] env[62460]: self.force_reraise() [ 658.732526] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 658.732526] env[62460]: raise self.value [ 658.732526] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 658.732526] env[62460]: updated_port = self._update_port( [ 658.732526] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 658.732526] env[62460]: _ensure_no_port_binding_failure(port) [ 658.732526] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 658.732526] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 658.732526] env[62460]: nova.exception.PortBindingFailed: Binding failed for port 8ecec905-4891-4a60-a063-ef5f68a1373b, please check neutron logs for more information. [ 658.732526] env[62460]: Removing descriptor: 19 [ 658.743648] env[62460]: DEBUG oslo_concurrency.lockutils [req-f5007b40-e61f-4075-8dbc-6ebf01e5d84b req-3fdede88-bec9-417c-a680-9de54cd39d12 service nova] Releasing lock "refresh_cache-a24776d9-9950-4c83-9641-9675cabd5fd3" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 658.743648] env[62460]: DEBUG nova.compute.manager [req-f5007b40-e61f-4075-8dbc-6ebf01e5d84b req-3fdede88-bec9-417c-a680-9de54cd39d12 service nova] [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] Received event network-vif-deleted-07d0ef21-dd97-42e1-8b07-67b1feade067 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 658.829318] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.928s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 658.830036] env[62460]: ERROR nova.compute.manager [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3373c77f-c40e-4658-9249-466f197b9940, please check neutron logs for more information. [ 658.830036] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] Traceback (most recent call last): [ 658.830036] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 658.830036] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] self.driver.spawn(context, instance, image_meta, [ 658.830036] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 658.830036] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 658.830036] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 658.830036] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] vm_ref = self.build_virtual_machine(instance, [ 658.830036] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 658.830036] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] vif_infos = vmwarevif.get_vif_info(self._session, [ 658.830036] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 658.830391] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] for vif in network_info: [ 658.830391] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 658.830391] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] return self._sync_wrapper(fn, *args, **kwargs) [ 658.830391] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 658.830391] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] self.wait() [ 658.830391] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 658.830391] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] self[:] = self._gt.wait() [ 658.830391] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 658.830391] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] return self._exit_event.wait() [ 658.830391] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 658.830391] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] result = hub.switch() [ 658.830391] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 658.830391] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] return self.greenlet.switch() [ 658.830696] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 658.830696] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] result = function(*args, **kwargs) [ 658.830696] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 658.830696] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] return func(*args, **kwargs) [ 658.830696] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 658.830696] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] raise e [ 658.830696] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 658.830696] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] nwinfo = self.network_api.allocate_for_instance( [ 658.830696] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 658.830696] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] created_port_ids = self._update_ports_for_instance( [ 658.830696] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 658.830696] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] with excutils.save_and_reraise_exception(): [ 658.830696] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 658.831120] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] self.force_reraise() [ 658.831120] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 658.831120] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] raise self.value [ 658.831120] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 658.831120] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] updated_port = self._update_port( [ 658.831120] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 658.831120] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] _ensure_no_port_binding_failure(port) [ 658.831120] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 658.831120] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] raise exception.PortBindingFailed(port_id=port['id']) [ 658.831120] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] nova.exception.PortBindingFailed: Binding failed for port 3373c77f-c40e-4658-9249-466f197b9940, please check neutron logs for more information. [ 658.831120] env[62460]: ERROR nova.compute.manager [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] [ 658.831409] env[62460]: DEBUG nova.compute.utils [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] Binding failed for port 3373c77f-c40e-4658-9249-466f197b9940, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 658.831520] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 28.240s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 658.835986] env[62460]: DEBUG nova.compute.manager [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] Build of instance 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6 was re-scheduled: Binding failed for port 3373c77f-c40e-4658-9249-466f197b9940, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 658.835986] env[62460]: DEBUG nova.compute.manager [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 658.835986] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "refresh_cache-9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 658.835986] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquired lock "refresh_cache-9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 658.835986] env[62460]: DEBUG nova.network.neutron [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 658.924740] env[62460]: DEBUG nova.compute.manager [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] [instance: 61426715-7a38-475d-895a-0eb6d6040c66] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 658.949310] env[62460]: DEBUG nova.virt.hardware [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 658.949623] env[62460]: DEBUG nova.virt.hardware [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 658.949801] env[62460]: DEBUG nova.virt.hardware [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 658.949983] env[62460]: DEBUG nova.virt.hardware [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 658.950167] env[62460]: DEBUG nova.virt.hardware [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 658.950288] env[62460]: DEBUG nova.virt.hardware [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 658.950496] env[62460]: DEBUG nova.virt.hardware [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 658.950655] env[62460]: DEBUG nova.virt.hardware [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 658.950817] env[62460]: DEBUG nova.virt.hardware [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 658.950978] env[62460]: DEBUG nova.virt.hardware [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 658.951163] env[62460]: DEBUG nova.virt.hardware [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 658.952082] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98586b48-0412-4095-a070-76004a25c370 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.960390] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a262865e-79c6-4035-9a51-436a41030e0f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.974367] env[62460]: ERROR nova.compute.manager [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] [instance: 61426715-7a38-475d-895a-0eb6d6040c66] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8ecec905-4891-4a60-a063-ef5f68a1373b, please check neutron logs for more information. [ 658.974367] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] Traceback (most recent call last): [ 658.974367] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 658.974367] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] yield resources [ 658.974367] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 658.974367] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] self.driver.spawn(context, instance, image_meta, [ 658.974367] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 658.974367] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] self._vmops.spawn(context, instance, image_meta, injected_files, [ 658.974367] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 658.974367] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] vm_ref = self.build_virtual_machine(instance, [ 658.974367] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 658.974716] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] vif_infos = vmwarevif.get_vif_info(self._session, [ 658.974716] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 658.974716] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] for vif in network_info: [ 658.974716] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 658.974716] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] return self._sync_wrapper(fn, *args, **kwargs) [ 658.974716] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 658.974716] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] self.wait() [ 658.974716] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 658.974716] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] self[:] = self._gt.wait() [ 658.974716] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 658.974716] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] return self._exit_event.wait() [ 658.974716] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 658.974716] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] current.throw(*self._exc) [ 658.975102] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 658.975102] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] result = function(*args, **kwargs) [ 658.975102] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 658.975102] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] return func(*args, **kwargs) [ 658.975102] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 658.975102] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] raise e [ 658.975102] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 658.975102] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] nwinfo = self.network_api.allocate_for_instance( [ 658.975102] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 658.975102] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] created_port_ids = self._update_ports_for_instance( [ 658.975102] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 658.975102] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] with excutils.save_and_reraise_exception(): [ 658.975102] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 658.975534] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] self.force_reraise() [ 658.975534] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 658.975534] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] raise self.value [ 658.975534] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 658.975534] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] updated_port = self._update_port( [ 658.975534] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 658.975534] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] _ensure_no_port_binding_failure(port) [ 658.975534] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 658.975534] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] raise exception.PortBindingFailed(port_id=port['id']) [ 658.975534] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] nova.exception.PortBindingFailed: Binding failed for port 8ecec905-4891-4a60-a063-ef5f68a1373b, please check neutron logs for more information. [ 658.975534] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] [ 658.975534] env[62460]: INFO nova.compute.manager [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] [instance: 61426715-7a38-475d-895a-0eb6d6040c66] Terminating instance [ 658.976674] env[62460]: DEBUG oslo_concurrency.lockutils [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] Acquiring lock "refresh_cache-61426715-7a38-475d-895a-0eb6d6040c66" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 658.976835] env[62460]: DEBUG oslo_concurrency.lockutils [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] Acquired lock "refresh_cache-61426715-7a38-475d-895a-0eb6d6040c66" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 658.976999] env[62460]: DEBUG nova.network.neutron [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] [instance: 61426715-7a38-475d-895a-0eb6d6040c66] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 659.358752] env[62460]: DEBUG nova.network.neutron [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 659.370716] env[62460]: DEBUG nova.compute.manager [req-837c4b34-adff-4522-aff1-2c7bd3f9dc7f req-1925bdb3-0540-4eeb-8006-9fca3c915777 service nova] [instance: 61426715-7a38-475d-895a-0eb6d6040c66] Received event network-changed-8ecec905-4891-4a60-a063-ef5f68a1373b {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 659.370944] env[62460]: DEBUG nova.compute.manager [req-837c4b34-adff-4522-aff1-2c7bd3f9dc7f req-1925bdb3-0540-4eeb-8006-9fca3c915777 service nova] [instance: 61426715-7a38-475d-895a-0eb6d6040c66] Refreshing instance network info cache due to event network-changed-8ecec905-4891-4a60-a063-ef5f68a1373b. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 659.371137] env[62460]: DEBUG oslo_concurrency.lockutils [req-837c4b34-adff-4522-aff1-2c7bd3f9dc7f req-1925bdb3-0540-4eeb-8006-9fca3c915777 service nova] Acquiring lock "refresh_cache-61426715-7a38-475d-895a-0eb6d6040c66" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 659.458630] env[62460]: DEBUG nova.network.neutron [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.494349] env[62460]: DEBUG nova.network.neutron [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] [instance: 61426715-7a38-475d-895a-0eb6d6040c66] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 659.574499] env[62460]: DEBUG nova.network.neutron [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] [instance: 61426715-7a38-475d-895a-0eb6d6040c66] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.647290] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78d1a0d2-03b0-4e9d-b854-73d413694cd4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.655289] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cbbed2d-5e84-4313-8806-b7cb231f6a95 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.686912] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bd10e5a-6303-4e95-8044-19cb0f0386ea {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.694200] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a99ab50f-ab4f-47aa-872d-1df858eac43b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.708365] env[62460]: DEBUG nova.compute.provider_tree [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 659.962075] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Releasing lock "refresh_cache-9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 659.962440] env[62460]: DEBUG nova.compute.manager [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 659.962440] env[62460]: DEBUG nova.compute.manager [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 659.962627] env[62460]: DEBUG nova.network.neutron [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 659.977516] env[62460]: DEBUG nova.network.neutron [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 660.077382] env[62460]: DEBUG oslo_concurrency.lockutils [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] Releasing lock "refresh_cache-61426715-7a38-475d-895a-0eb6d6040c66" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 660.077821] env[62460]: DEBUG nova.compute.manager [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] [instance: 61426715-7a38-475d-895a-0eb6d6040c66] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 660.078030] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] [instance: 61426715-7a38-475d-895a-0eb6d6040c66] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 660.078343] env[62460]: DEBUG oslo_concurrency.lockutils [req-837c4b34-adff-4522-aff1-2c7bd3f9dc7f req-1925bdb3-0540-4eeb-8006-9fca3c915777 service nova] Acquired lock "refresh_cache-61426715-7a38-475d-895a-0eb6d6040c66" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 660.078521] env[62460]: DEBUG nova.network.neutron [req-837c4b34-adff-4522-aff1-2c7bd3f9dc7f req-1925bdb3-0540-4eeb-8006-9fca3c915777 service nova] [instance: 61426715-7a38-475d-895a-0eb6d6040c66] Refreshing network info cache for port 8ecec905-4891-4a60-a063-ef5f68a1373b {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 660.079571] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-95b390d6-f369-4854-abfb-ddf6b84ae504 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.089140] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca7bc097-c410-4345-8d3b-4d72834161e0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.110983] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] [instance: 61426715-7a38-475d-895a-0eb6d6040c66] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 61426715-7a38-475d-895a-0eb6d6040c66 could not be found. [ 660.110983] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] [instance: 61426715-7a38-475d-895a-0eb6d6040c66] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 660.110983] env[62460]: INFO nova.compute.manager [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] [instance: 61426715-7a38-475d-895a-0eb6d6040c66] Took 0.03 seconds to destroy the instance on the hypervisor. [ 660.110983] env[62460]: DEBUG oslo.service.loopingcall [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 660.112258] env[62460]: DEBUG nova.compute.manager [-] [instance: 61426715-7a38-475d-895a-0eb6d6040c66] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 660.112258] env[62460]: DEBUG nova.network.neutron [-] [instance: 61426715-7a38-475d-895a-0eb6d6040c66] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 660.125126] env[62460]: DEBUG nova.network.neutron [-] [instance: 61426715-7a38-475d-895a-0eb6d6040c66] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 660.211923] env[62460]: DEBUG nova.scheduler.client.report [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 660.480508] env[62460]: DEBUG nova.network.neutron [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 660.596933] env[62460]: DEBUG nova.network.neutron [req-837c4b34-adff-4522-aff1-2c7bd3f9dc7f req-1925bdb3-0540-4eeb-8006-9fca3c915777 service nova] [instance: 61426715-7a38-475d-895a-0eb6d6040c66] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 660.627331] env[62460]: DEBUG nova.network.neutron [-] [instance: 61426715-7a38-475d-895a-0eb6d6040c66] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 660.675963] env[62460]: DEBUG nova.network.neutron [req-837c4b34-adff-4522-aff1-2c7bd3f9dc7f req-1925bdb3-0540-4eeb-8006-9fca3c915777 service nova] [instance: 61426715-7a38-475d-895a-0eb6d6040c66] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 660.716833] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.885s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 660.717513] env[62460]: ERROR nova.compute.manager [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3f617afc-843a-4ab7-88fe-7a981af30004, please check neutron logs for more information. [ 660.717513] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] Traceback (most recent call last): [ 660.717513] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 660.717513] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] self.driver.spawn(context, instance, image_meta, [ 660.717513] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 660.717513] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] self._vmops.spawn(context, instance, image_meta, injected_files, [ 660.717513] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 660.717513] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] vm_ref = self.build_virtual_machine(instance, [ 660.717513] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 660.717513] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] vif_infos = vmwarevif.get_vif_info(self._session, [ 660.717513] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 660.717834] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] for vif in network_info: [ 660.717834] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 660.717834] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] return self._sync_wrapper(fn, *args, **kwargs) [ 660.717834] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 660.717834] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] self.wait() [ 660.717834] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 660.717834] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] self[:] = self._gt.wait() [ 660.717834] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 660.717834] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] return self._exit_event.wait() [ 660.717834] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 660.717834] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] result = hub.switch() [ 660.717834] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 660.717834] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] return self.greenlet.switch() [ 660.718210] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 660.718210] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] result = function(*args, **kwargs) [ 660.718210] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 660.718210] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] return func(*args, **kwargs) [ 660.718210] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 660.718210] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] raise e [ 660.718210] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 660.718210] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] nwinfo = self.network_api.allocate_for_instance( [ 660.718210] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 660.718210] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] created_port_ids = self._update_ports_for_instance( [ 660.718210] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 660.718210] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] with excutils.save_and_reraise_exception(): [ 660.718210] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 660.718583] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] self.force_reraise() [ 660.718583] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 660.718583] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] raise self.value [ 660.718583] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 660.718583] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] updated_port = self._update_port( [ 660.718583] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 660.718583] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] _ensure_no_port_binding_failure(port) [ 660.718583] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 660.718583] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] raise exception.PortBindingFailed(port_id=port['id']) [ 660.718583] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] nova.exception.PortBindingFailed: Binding failed for port 3f617afc-843a-4ab7-88fe-7a981af30004, please check neutron logs for more information. [ 660.718583] env[62460]: ERROR nova.compute.manager [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] [ 660.718903] env[62460]: DEBUG nova.compute.utils [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] Binding failed for port 3f617afc-843a-4ab7-88fe-7a981af30004, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 660.719556] env[62460]: DEBUG oslo_concurrency.lockutils [None req-398cf1be-cafa-4d72-b54f-722129802cf2 tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.363s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 660.719776] env[62460]: DEBUG oslo_concurrency.lockutils [None req-398cf1be-cafa-4d72-b54f-722129802cf2 tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 660.722036] env[62460]: DEBUG oslo_concurrency.lockutils [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.112s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 660.723720] env[62460]: INFO nova.compute.claims [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 660.728069] env[62460]: DEBUG nova.compute.manager [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] Build of instance 01bbc9ec-f2ee-4a48-a33c-784861e81097 was re-scheduled: Binding failed for port 3f617afc-843a-4ab7-88fe-7a981af30004, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 660.728529] env[62460]: DEBUG nova.compute.manager [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 660.728756] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Acquiring lock "refresh_cache-01bbc9ec-f2ee-4a48-a33c-784861e81097" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 660.728903] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Acquired lock "refresh_cache-01bbc9ec-f2ee-4a48-a33c-784861e81097" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 660.729077] env[62460]: DEBUG nova.network.neutron [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 660.747234] env[62460]: INFO nova.scheduler.client.report [None req-398cf1be-cafa-4d72-b54f-722129802cf2 tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Deleted allocations for instance 3b71c366-cbd9-4b98-aa0d-c55b56e69231 [ 660.982941] env[62460]: INFO nova.compute.manager [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6] Took 1.02 seconds to deallocate network for instance. [ 661.130066] env[62460]: INFO nova.compute.manager [-] [instance: 61426715-7a38-475d-895a-0eb6d6040c66] Took 1.02 seconds to deallocate network for instance. [ 661.132950] env[62460]: DEBUG nova.compute.claims [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] [instance: 61426715-7a38-475d-895a-0eb6d6040c66] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 661.133129] env[62460]: DEBUG oslo_concurrency.lockutils [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 661.178366] env[62460]: DEBUG oslo_concurrency.lockutils [req-837c4b34-adff-4522-aff1-2c7bd3f9dc7f req-1925bdb3-0540-4eeb-8006-9fca3c915777 service nova] Releasing lock "refresh_cache-61426715-7a38-475d-895a-0eb6d6040c66" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 661.178652] env[62460]: DEBUG nova.compute.manager [req-837c4b34-adff-4522-aff1-2c7bd3f9dc7f req-1925bdb3-0540-4eeb-8006-9fca3c915777 service nova] [instance: 61426715-7a38-475d-895a-0eb6d6040c66] Received event network-vif-deleted-8ecec905-4891-4a60-a063-ef5f68a1373b {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 661.247669] env[62460]: DEBUG nova.network.neutron [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 661.253244] env[62460]: DEBUG oslo_concurrency.lockutils [None req-398cf1be-cafa-4d72-b54f-722129802cf2 tempest-ServersAaction247Test-422733556 tempest-ServersAaction247Test-422733556-project-member] Lock "3b71c366-cbd9-4b98-aa0d-c55b56e69231" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.697s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 661.315874] env[62460]: DEBUG nova.network.neutron [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 661.817845] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Releasing lock "refresh_cache-01bbc9ec-f2ee-4a48-a33c-784861e81097" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 661.818118] env[62460]: DEBUG nova.compute.manager [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 661.818332] env[62460]: DEBUG nova.compute.manager [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 661.818530] env[62460]: DEBUG nova.network.neutron [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 661.841011] env[62460]: DEBUG nova.network.neutron [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 662.017569] env[62460]: INFO nova.scheduler.client.report [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Deleted allocations for instance 9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6 [ 662.118830] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eadeb414-78d3-45af-96eb-98cfaa532e03 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.126278] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85233df5-856e-413b-aea2-f88e3f084748 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.157307] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f1dd759-66f3-4fdf-85d0-23ae903c550d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.165064] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e4afdd3-955f-4c03-af5e-bba006fb539f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.178718] env[62460]: DEBUG nova.compute.provider_tree [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 662.344979] env[62460]: DEBUG nova.network.neutron [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.532284] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e3bcdc25-ad5c-47d4-a097-cf3e631d1555 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "9c7a44b1-d87a-42a7-95f3-b0b064ce1ef6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 98.504s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 662.681493] env[62460]: DEBUG nova.scheduler.client.report [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 662.847323] env[62460]: INFO nova.compute.manager [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: 01bbc9ec-f2ee-4a48-a33c-784861e81097] Took 1.03 seconds to deallocate network for instance. [ 663.034845] env[62460]: DEBUG nova.compute.manager [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 663.187054] env[62460]: DEBUG oslo_concurrency.lockutils [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.465s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 663.187603] env[62460]: DEBUG nova.compute.manager [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 663.190848] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 28.209s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 663.553304] env[62460]: DEBUG oslo_concurrency.lockutils [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 663.695808] env[62460]: DEBUG nova.compute.utils [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 663.699936] env[62460]: DEBUG nova.compute.manager [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 663.700125] env[62460]: DEBUG nova.network.neutron [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 663.746711] env[62460]: DEBUG nova.policy [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd7b38fb0a53f4d078ae86d68e3eb6268', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '53fa06008e9f43488362895e7a143700', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 663.894204] env[62460]: INFO nova.scheduler.client.report [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Deleted allocations for instance 01bbc9ec-f2ee-4a48-a33c-784861e81097 [ 664.123304] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4203e51-145f-48f3-95b8-a43c0f083532 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.126784] env[62460]: DEBUG nova.network.neutron [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] Successfully created port: 58dfc745-c39d-4558-af1a-1f04af4a3b87 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 664.132935] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dd875e9-e880-46d9-89b9-8b3bd9cc24f5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.164124] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-066bedbd-9d76-4ca7-97fc-6e8080ebc40f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.172184] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab503739-6678-4793-8084-d31e1d8d541a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.185844] env[62460]: DEBUG nova.compute.provider_tree [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 664.203516] env[62460]: DEBUG nova.compute.manager [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 664.317076] env[62460]: DEBUG oslo_concurrency.lockutils [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "e472b96a-5d82-4fb9-8d28-0901b71dc783" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 664.317329] env[62460]: DEBUG oslo_concurrency.lockutils [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "e472b96a-5d82-4fb9-8d28-0901b71dc783" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 664.405734] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c018472e-0ed2-40c6-a828-a260bd20d962 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Lock "01bbc9ec-f2ee-4a48-a33c-784861e81097" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 99.960s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 664.692018] env[62460]: DEBUG nova.scheduler.client.report [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 664.876330] env[62460]: DEBUG nova.compute.manager [req-64aacd5e-8724-438a-83c9-a231909e5fc2 req-ec5070c7-698c-44a4-b0bb-6bfdfca6d780 service nova] [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] Received event network-changed-58dfc745-c39d-4558-af1a-1f04af4a3b87 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 664.876330] env[62460]: DEBUG nova.compute.manager [req-64aacd5e-8724-438a-83c9-a231909e5fc2 req-ec5070c7-698c-44a4-b0bb-6bfdfca6d780 service nova] [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] Refreshing instance network info cache due to event network-changed-58dfc745-c39d-4558-af1a-1f04af4a3b87. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 664.876330] env[62460]: DEBUG oslo_concurrency.lockutils [req-64aacd5e-8724-438a-83c9-a231909e5fc2 req-ec5070c7-698c-44a4-b0bb-6bfdfca6d780 service nova] Acquiring lock "refresh_cache-b6efad50-aa2e-49f9-9ce7-5fead31db7a8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 664.876330] env[62460]: DEBUG oslo_concurrency.lockutils [req-64aacd5e-8724-438a-83c9-a231909e5fc2 req-ec5070c7-698c-44a4-b0bb-6bfdfca6d780 service nova] Acquired lock "refresh_cache-b6efad50-aa2e-49f9-9ce7-5fead31db7a8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 664.876330] env[62460]: DEBUG nova.network.neutron [req-64aacd5e-8724-438a-83c9-a231909e5fc2 req-ec5070c7-698c-44a4-b0bb-6bfdfca6d780 service nova] [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] Refreshing network info cache for port 58dfc745-c39d-4558-af1a-1f04af4a3b87 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 664.909516] env[62460]: DEBUG nova.compute.manager [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 665.101610] env[62460]: ERROR nova.compute.manager [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 58dfc745-c39d-4558-af1a-1f04af4a3b87, please check neutron logs for more information. [ 665.101610] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 665.101610] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 665.101610] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 665.101610] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 665.101610] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 665.101610] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 665.101610] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 665.101610] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 665.101610] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 665.101610] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 665.101610] env[62460]: ERROR nova.compute.manager raise self.value [ 665.101610] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 665.101610] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 665.101610] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 665.101610] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 665.102096] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 665.102096] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 665.102096] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 58dfc745-c39d-4558-af1a-1f04af4a3b87, please check neutron logs for more information. [ 665.102096] env[62460]: ERROR nova.compute.manager [ 665.102096] env[62460]: Traceback (most recent call last): [ 665.102096] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 665.102096] env[62460]: listener.cb(fileno) [ 665.102096] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 665.102096] env[62460]: result = function(*args, **kwargs) [ 665.102096] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 665.102096] env[62460]: return func(*args, **kwargs) [ 665.102096] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 665.102096] env[62460]: raise e [ 665.102096] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 665.102096] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 665.102096] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 665.102096] env[62460]: created_port_ids = self._update_ports_for_instance( [ 665.102096] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 665.102096] env[62460]: with excutils.save_and_reraise_exception(): [ 665.102096] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 665.102096] env[62460]: self.force_reraise() [ 665.102096] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 665.102096] env[62460]: raise self.value [ 665.102096] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 665.102096] env[62460]: updated_port = self._update_port( [ 665.102096] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 665.102096] env[62460]: _ensure_no_port_binding_failure(port) [ 665.102096] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 665.102096] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 665.103044] env[62460]: nova.exception.PortBindingFailed: Binding failed for port 58dfc745-c39d-4558-af1a-1f04af4a3b87, please check neutron logs for more information. [ 665.103044] env[62460]: Removing descriptor: 19 [ 665.196644] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.004s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 665.196644] env[62460]: ERROR nova.compute.manager [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] [instance: 94703126-bb31-40af-b945-2ef5ff37e094] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 620a0798-9892-49ff-9a8c-a60a330f5b35, please check neutron logs for more information. [ 665.196644] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] Traceback (most recent call last): [ 665.196644] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 665.196644] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] self.driver.spawn(context, instance, image_meta, [ 665.196644] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 665.196644] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] self._vmops.spawn(context, instance, image_meta, injected_files, [ 665.196644] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 665.196644] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] vm_ref = self.build_virtual_machine(instance, [ 665.197123] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 665.197123] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] vif_infos = vmwarevif.get_vif_info(self._session, [ 665.197123] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 665.197123] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] for vif in network_info: [ 665.197123] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 665.197123] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] return self._sync_wrapper(fn, *args, **kwargs) [ 665.197123] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 665.197123] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] self.wait() [ 665.197123] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 665.197123] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] self[:] = self._gt.wait() [ 665.197123] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 665.197123] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] return self._exit_event.wait() [ 665.197123] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 665.197483] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] result = hub.switch() [ 665.197483] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 665.197483] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] return self.greenlet.switch() [ 665.197483] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 665.197483] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] result = function(*args, **kwargs) [ 665.197483] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 665.197483] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] return func(*args, **kwargs) [ 665.197483] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 665.197483] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] raise e [ 665.197483] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 665.197483] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] nwinfo = self.network_api.allocate_for_instance( [ 665.197483] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 665.197483] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] created_port_ids = self._update_ports_for_instance( [ 665.197859] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 665.197859] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] with excutils.save_and_reraise_exception(): [ 665.197859] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 665.197859] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] self.force_reraise() [ 665.197859] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 665.197859] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] raise self.value [ 665.197859] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 665.197859] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] updated_port = self._update_port( [ 665.197859] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 665.197859] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] _ensure_no_port_binding_failure(port) [ 665.197859] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 665.197859] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] raise exception.PortBindingFailed(port_id=port['id']) [ 665.198196] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] nova.exception.PortBindingFailed: Binding failed for port 620a0798-9892-49ff-9a8c-a60a330f5b35, please check neutron logs for more information. [ 665.198196] env[62460]: ERROR nova.compute.manager [instance: 94703126-bb31-40af-b945-2ef5ff37e094] [ 665.198196] env[62460]: DEBUG nova.compute.utils [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] [instance: 94703126-bb31-40af-b945-2ef5ff37e094] Binding failed for port 620a0798-9892-49ff-9a8c-a60a330f5b35, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 665.199346] env[62460]: DEBUG oslo_concurrency.lockutils [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 28.096s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 665.203147] env[62460]: DEBUG nova.compute.manager [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] [instance: 94703126-bb31-40af-b945-2ef5ff37e094] Build of instance 94703126-bb31-40af-b945-2ef5ff37e094 was re-scheduled: Binding failed for port 620a0798-9892-49ff-9a8c-a60a330f5b35, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 665.203592] env[62460]: DEBUG nova.compute.manager [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] [instance: 94703126-bb31-40af-b945-2ef5ff37e094] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 665.203824] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] Acquiring lock "refresh_cache-94703126-bb31-40af-b945-2ef5ff37e094" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 665.205594] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] Acquired lock "refresh_cache-94703126-bb31-40af-b945-2ef5ff37e094" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 665.205594] env[62460]: DEBUG nova.network.neutron [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] [instance: 94703126-bb31-40af-b945-2ef5ff37e094] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 665.212078] env[62460]: DEBUG nova.compute.manager [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 665.240179] env[62460]: DEBUG nova.virt.hardware [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 665.240433] env[62460]: DEBUG nova.virt.hardware [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 665.240593] env[62460]: DEBUG nova.virt.hardware [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 665.240776] env[62460]: DEBUG nova.virt.hardware [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 665.240924] env[62460]: DEBUG nova.virt.hardware [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 665.241088] env[62460]: DEBUG nova.virt.hardware [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 665.241299] env[62460]: DEBUG nova.virt.hardware [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 665.241952] env[62460]: DEBUG nova.virt.hardware [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 665.241952] env[62460]: DEBUG nova.virt.hardware [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 665.241952] env[62460]: DEBUG nova.virt.hardware [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 665.242103] env[62460]: DEBUG nova.virt.hardware [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 665.242950] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fa8d27b-a941-46eb-96d2-b0b2b3f11d68 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.251578] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f5d432f-360a-4651-bc08-31ab9f7182b2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.266231] env[62460]: ERROR nova.compute.manager [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 58dfc745-c39d-4558-af1a-1f04af4a3b87, please check neutron logs for more information. [ 665.266231] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] Traceback (most recent call last): [ 665.266231] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 665.266231] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] yield resources [ 665.266231] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 665.266231] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] self.driver.spawn(context, instance, image_meta, [ 665.266231] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 665.266231] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 665.266231] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 665.266231] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] vm_ref = self.build_virtual_machine(instance, [ 665.266231] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 665.266606] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] vif_infos = vmwarevif.get_vif_info(self._session, [ 665.266606] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 665.266606] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] for vif in network_info: [ 665.266606] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 665.266606] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] return self._sync_wrapper(fn, *args, **kwargs) [ 665.266606] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 665.266606] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] self.wait() [ 665.266606] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 665.266606] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] self[:] = self._gt.wait() [ 665.266606] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 665.266606] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] return self._exit_event.wait() [ 665.266606] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 665.266606] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] current.throw(*self._exc) [ 665.266945] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 665.266945] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] result = function(*args, **kwargs) [ 665.266945] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 665.266945] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] return func(*args, **kwargs) [ 665.266945] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 665.266945] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] raise e [ 665.266945] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 665.266945] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] nwinfo = self.network_api.allocate_for_instance( [ 665.266945] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 665.266945] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] created_port_ids = self._update_ports_for_instance( [ 665.266945] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 665.266945] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] with excutils.save_and_reraise_exception(): [ 665.266945] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 665.267283] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] self.force_reraise() [ 665.267283] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 665.267283] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] raise self.value [ 665.267283] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 665.267283] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] updated_port = self._update_port( [ 665.267283] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 665.267283] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] _ensure_no_port_binding_failure(port) [ 665.267283] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 665.267283] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] raise exception.PortBindingFailed(port_id=port['id']) [ 665.267283] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] nova.exception.PortBindingFailed: Binding failed for port 58dfc745-c39d-4558-af1a-1f04af4a3b87, please check neutron logs for more information. [ 665.267283] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] [ 665.267283] env[62460]: INFO nova.compute.manager [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] Terminating instance [ 665.268484] env[62460]: DEBUG oslo_concurrency.lockutils [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "refresh_cache-b6efad50-aa2e-49f9-9ce7-5fead31db7a8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 665.397278] env[62460]: DEBUG nova.network.neutron [req-64aacd5e-8724-438a-83c9-a231909e5fc2 req-ec5070c7-698c-44a4-b0bb-6bfdfca6d780 service nova] [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 665.434430] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 665.519982] env[62460]: DEBUG nova.network.neutron [req-64aacd5e-8724-438a-83c9-a231909e5fc2 req-ec5070c7-698c-44a4-b0bb-6bfdfca6d780 service nova] [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.735124] env[62460]: DEBUG nova.network.neutron [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] [instance: 94703126-bb31-40af-b945-2ef5ff37e094] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 665.798288] env[62460]: DEBUG nova.network.neutron [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] [instance: 94703126-bb31-40af-b945-2ef5ff37e094] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 666.022502] env[62460]: DEBUG oslo_concurrency.lockutils [req-64aacd5e-8724-438a-83c9-a231909e5fc2 req-ec5070c7-698c-44a4-b0bb-6bfdfca6d780 service nova] Releasing lock "refresh_cache-b6efad50-aa2e-49f9-9ce7-5fead31db7a8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 666.022949] env[62460]: DEBUG oslo_concurrency.lockutils [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquired lock "refresh_cache-b6efad50-aa2e-49f9-9ce7-5fead31db7a8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 666.023148] env[62460]: DEBUG nova.network.neutron [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 666.036359] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7fb09b9-2852-46c7-91b4-515c4a8b1791 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.043990] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fdf365d-cc21-4fa5-aee4-ea636e9b3c20 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.074950] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29184b15-b316-4ca6-a057-10b991bb789f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.082784] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4a3c4a3-1500-4df1-982a-54209303193d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.096300] env[62460]: DEBUG nova.compute.provider_tree [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 666.257148] env[62460]: DEBUG oslo_concurrency.lockutils [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Acquiring lock "be50b542-4c61-4da3-af4e-8a57e05a77ee" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 666.257402] env[62460]: DEBUG oslo_concurrency.lockutils [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Lock "be50b542-4c61-4da3-af4e-8a57e05a77ee" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 666.302623] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] Releasing lock "refresh_cache-94703126-bb31-40af-b945-2ef5ff37e094" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 666.302890] env[62460]: DEBUG nova.compute.manager [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 666.303084] env[62460]: DEBUG nova.compute.manager [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] [instance: 94703126-bb31-40af-b945-2ef5ff37e094] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 666.303264] env[62460]: DEBUG nova.network.neutron [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] [instance: 94703126-bb31-40af-b945-2ef5ff37e094] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 666.318213] env[62460]: DEBUG nova.network.neutron [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] [instance: 94703126-bb31-40af-b945-2ef5ff37e094] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 666.539788] env[62460]: DEBUG nova.network.neutron [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 666.599108] env[62460]: DEBUG nova.scheduler.client.report [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 666.636108] env[62460]: DEBUG nova.network.neutron [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 666.821170] env[62460]: DEBUG nova.network.neutron [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] [instance: 94703126-bb31-40af-b945-2ef5ff37e094] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 666.899034] env[62460]: DEBUG nova.compute.manager [req-a4192d57-5256-4371-9924-ead1d1c65656 req-d6ee30b7-571f-46bc-88d1-02f413edddba service nova] [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] Received event network-vif-deleted-58dfc745-c39d-4558-af1a-1f04af4a3b87 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 667.104349] env[62460]: DEBUG oslo_concurrency.lockutils [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.905s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 667.105034] env[62460]: ERROR nova.compute.manager [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 10776c82-fe5d-46e6-b9e6-7aacb6b11ab9, please check neutron logs for more information. [ 667.105034] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] Traceback (most recent call last): [ 667.105034] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 667.105034] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] self.driver.spawn(context, instance, image_meta, [ 667.105034] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 667.105034] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 667.105034] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 667.105034] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] vm_ref = self.build_virtual_machine(instance, [ 667.105034] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 667.105034] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] vif_infos = vmwarevif.get_vif_info(self._session, [ 667.105034] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 667.105403] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] for vif in network_info: [ 667.105403] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 667.105403] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] return self._sync_wrapper(fn, *args, **kwargs) [ 667.105403] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 667.105403] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] self.wait() [ 667.105403] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 667.105403] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] self[:] = self._gt.wait() [ 667.105403] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 667.105403] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] return self._exit_event.wait() [ 667.105403] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 667.105403] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] current.throw(*self._exc) [ 667.105403] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 667.105403] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] result = function(*args, **kwargs) [ 667.105781] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 667.105781] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] return func(*args, **kwargs) [ 667.105781] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 667.105781] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] raise e [ 667.105781] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 667.105781] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] nwinfo = self.network_api.allocate_for_instance( [ 667.105781] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 667.105781] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] created_port_ids = self._update_ports_for_instance( [ 667.105781] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 667.105781] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] with excutils.save_and_reraise_exception(): [ 667.105781] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 667.105781] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] self.force_reraise() [ 667.105781] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 667.106156] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] raise self.value [ 667.106156] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 667.106156] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] updated_port = self._update_port( [ 667.106156] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 667.106156] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] _ensure_no_port_binding_failure(port) [ 667.106156] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 667.106156] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] raise exception.PortBindingFailed(port_id=port['id']) [ 667.106156] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] nova.exception.PortBindingFailed: Binding failed for port 10776c82-fe5d-46e6-b9e6-7aacb6b11ab9, please check neutron logs for more information. [ 667.106156] env[62460]: ERROR nova.compute.manager [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] [ 667.106156] env[62460]: DEBUG nova.compute.utils [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] Binding failed for port 10776c82-fe5d-46e6-b9e6-7aacb6b11ab9, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 667.107326] env[62460]: DEBUG oslo_concurrency.lockutils [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 28.592s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 667.110141] env[62460]: DEBUG nova.compute.manager [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] Build of instance 3426ace3-0cff-4119-85fe-31e681a8597b was re-scheduled: Binding failed for port 10776c82-fe5d-46e6-b9e6-7aacb6b11ab9, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 667.110870] env[62460]: DEBUG nova.compute.manager [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 667.110870] env[62460]: DEBUG oslo_concurrency.lockutils [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] Acquiring lock "refresh_cache-3426ace3-0cff-4119-85fe-31e681a8597b" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 667.110987] env[62460]: DEBUG oslo_concurrency.lockutils [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] Acquired lock "refresh_cache-3426ace3-0cff-4119-85fe-31e681a8597b" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 667.111081] env[62460]: DEBUG nova.network.neutron [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 667.138640] env[62460]: DEBUG oslo_concurrency.lockutils [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Releasing lock "refresh_cache-b6efad50-aa2e-49f9-9ce7-5fead31db7a8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 667.138961] env[62460]: DEBUG nova.compute.manager [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 667.139204] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 667.139527] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0409df29-e276-4846-82de-503696360b04 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.148468] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62d125c0-ffda-4a71-a728-d52db8c17dc9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.170487] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b6efad50-aa2e-49f9-9ce7-5fead31db7a8 could not be found. [ 667.170698] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 667.170879] env[62460]: INFO nova.compute.manager [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] Took 0.03 seconds to destroy the instance on the hypervisor. [ 667.171132] env[62460]: DEBUG oslo.service.loopingcall [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 667.171345] env[62460]: DEBUG nova.compute.manager [-] [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 667.171441] env[62460]: DEBUG nova.network.neutron [-] [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 667.185233] env[62460]: DEBUG nova.network.neutron [-] [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 667.323400] env[62460]: INFO nova.compute.manager [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] [instance: 94703126-bb31-40af-b945-2ef5ff37e094] Took 1.02 seconds to deallocate network for instance. [ 667.632937] env[62460]: DEBUG nova.network.neutron [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 667.688021] env[62460]: DEBUG nova.network.neutron [-] [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 667.718304] env[62460]: DEBUG nova.network.neutron [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 667.911246] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afeb8b7c-0309-4d3e-8cd1-b61d7cf3629c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.918636] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48ed7c3e-a68e-4ac1-a91e-805f5bb8bc8c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.947381] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2af295a6-35a0-4132-836e-d846cb41b0d9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.954895] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5c6f89b-7aac-4575-92ed-e37341e3ee95 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.969226] env[62460]: DEBUG nova.compute.provider_tree [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 668.189250] env[62460]: INFO nova.compute.manager [-] [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] Took 1.02 seconds to deallocate network for instance. [ 668.191945] env[62460]: DEBUG nova.compute.claims [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 668.192095] env[62460]: DEBUG oslo_concurrency.lockutils [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.222767] env[62460]: DEBUG oslo_concurrency.lockutils [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] Releasing lock "refresh_cache-3426ace3-0cff-4119-85fe-31e681a8597b" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 668.223010] env[62460]: DEBUG nova.compute.manager [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 668.223205] env[62460]: DEBUG nova.compute.manager [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 668.223376] env[62460]: DEBUG nova.network.neutron [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 668.237550] env[62460]: DEBUG nova.network.neutron [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 668.352448] env[62460]: INFO nova.scheduler.client.report [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] Deleted allocations for instance 94703126-bb31-40af-b945-2ef5ff37e094 [ 668.471956] env[62460]: DEBUG nova.scheduler.client.report [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 668.739862] env[62460]: DEBUG nova.network.neutron [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 668.860994] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0698fc3f-274c-460b-bd26-99e8ebc1d5b8 tempest-ServerDiagnosticsTest-20263120 tempest-ServerDiagnosticsTest-20263120-project-member] Lock "94703126-bb31-40af-b945-2ef5ff37e094" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 101.286s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 668.976935] env[62460]: DEBUG oslo_concurrency.lockutils [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.869s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 668.978633] env[62460]: ERROR nova.compute.manager [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3221cf96-ce6e-40d8-bb3b-7ed671e0b4d1, please check neutron logs for more information. [ 668.978633] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] Traceback (most recent call last): [ 668.978633] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 668.978633] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] self.driver.spawn(context, instance, image_meta, [ 668.978633] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 668.978633] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 668.978633] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 668.978633] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] vm_ref = self.build_virtual_machine(instance, [ 668.978633] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 668.978633] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] vif_infos = vmwarevif.get_vif_info(self._session, [ 668.978633] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 668.979133] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] for vif in network_info: [ 668.979133] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 668.979133] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] return self._sync_wrapper(fn, *args, **kwargs) [ 668.979133] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 668.979133] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] self.wait() [ 668.979133] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 668.979133] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] self[:] = self._gt.wait() [ 668.979133] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 668.979133] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] return self._exit_event.wait() [ 668.979133] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 668.979133] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] current.throw(*self._exc) [ 668.979133] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 668.979133] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] result = function(*args, **kwargs) [ 668.979520] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 668.979520] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] return func(*args, **kwargs) [ 668.979520] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 668.979520] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] raise e [ 668.979520] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 668.979520] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] nwinfo = self.network_api.allocate_for_instance( [ 668.979520] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 668.979520] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] created_port_ids = self._update_ports_for_instance( [ 668.979520] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 668.979520] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] with excutils.save_and_reraise_exception(): [ 668.979520] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 668.979520] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] self.force_reraise() [ 668.979520] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 668.980023] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] raise self.value [ 668.980023] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 668.980023] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] updated_port = self._update_port( [ 668.980023] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 668.980023] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] _ensure_no_port_binding_failure(port) [ 668.980023] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 668.980023] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] raise exception.PortBindingFailed(port_id=port['id']) [ 668.980023] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] nova.exception.PortBindingFailed: Binding failed for port 3221cf96-ce6e-40d8-bb3b-7ed671e0b4d1, please check neutron logs for more information. [ 668.980023] env[62460]: ERROR nova.compute.manager [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] [ 668.980023] env[62460]: DEBUG nova.compute.utils [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] Binding failed for port 3221cf96-ce6e-40d8-bb3b-7ed671e0b4d1, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 668.980323] env[62460]: DEBUG oslo_concurrency.lockutils [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.423s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 668.982913] env[62460]: DEBUG nova.compute.manager [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] Build of instance dba18c87-50a2-4ac8-8bd2-77f8f62b723c was re-scheduled: Binding failed for port 3221cf96-ce6e-40d8-bb3b-7ed671e0b4d1, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 668.983485] env[62460]: DEBUG nova.compute.manager [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 668.983568] env[62460]: DEBUG oslo_concurrency.lockutils [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Acquiring lock "refresh_cache-dba18c87-50a2-4ac8-8bd2-77f8f62b723c" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 668.983718] env[62460]: DEBUG oslo_concurrency.lockutils [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Acquired lock "refresh_cache-dba18c87-50a2-4ac8-8bd2-77f8f62b723c" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 668.983940] env[62460]: DEBUG nova.network.neutron [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 669.244090] env[62460]: INFO nova.compute.manager [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] [instance: 3426ace3-0cff-4119-85fe-31e681a8597b] Took 1.02 seconds to deallocate network for instance. [ 669.362155] env[62460]: DEBUG nova.compute.manager [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] [instance: 578627be-e695-4953-8d0e-9763d12b9a28] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 669.509356] env[62460]: DEBUG nova.network.neutron [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 669.608254] env[62460]: DEBUG nova.network.neutron [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 669.882736] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 669.889653] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caeced4f-5010-47ec-a772-c5bd9765dbc4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.897431] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13aa7f70-feda-4e95-b9fa-ce4267474ba0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.928920] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a88a56c-cce7-4644-a234-c8b71f8d91b7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.936410] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7afda0a1-5592-47ef-bcc1-f86ac24dd431 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.949546] env[62460]: DEBUG nova.compute.provider_tree [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 670.113735] env[62460]: DEBUG oslo_concurrency.lockutils [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Releasing lock "refresh_cache-dba18c87-50a2-4ac8-8bd2-77f8f62b723c" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 670.114054] env[62460]: DEBUG nova.compute.manager [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 670.114246] env[62460]: DEBUG nova.compute.manager [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 670.114417] env[62460]: DEBUG nova.network.neutron [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 670.129959] env[62460]: DEBUG nova.network.neutron [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 670.279089] env[62460]: INFO nova.scheduler.client.report [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] Deleted allocations for instance 3426ace3-0cff-4119-85fe-31e681a8597b [ 670.454517] env[62460]: DEBUG nova.scheduler.client.report [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 670.632360] env[62460]: DEBUG nova.network.neutron [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.787110] env[62460]: DEBUG oslo_concurrency.lockutils [None req-f08cb7d5-92e9-42be-8fce-75b6e9e0f025 tempest-ServersTestFqdnHostnames-593307524 tempest-ServersTestFqdnHostnames-593307524-project-member] Lock "3426ace3-0cff-4119-85fe-31e681a8597b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 102.281s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 670.959681] env[62460]: DEBUG oslo_concurrency.lockutils [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.980s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 670.964239] env[62460]: ERROR nova.compute.manager [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 759c443f-21d6-4fef-a27b-6fe32cc20a63, please check neutron logs for more information. [ 670.964239] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] Traceback (most recent call last): [ 670.964239] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 670.964239] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] self.driver.spawn(context, instance, image_meta, [ 670.964239] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 670.964239] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 670.964239] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 670.964239] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] vm_ref = self.build_virtual_machine(instance, [ 670.964239] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 670.964239] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] vif_infos = vmwarevif.get_vif_info(self._session, [ 670.964239] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 670.964822] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] for vif in network_info: [ 670.964822] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 670.964822] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] return self._sync_wrapper(fn, *args, **kwargs) [ 670.964822] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 670.964822] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] self.wait() [ 670.964822] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 670.964822] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] self[:] = self._gt.wait() [ 670.964822] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 670.964822] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] return self._exit_event.wait() [ 670.964822] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 670.964822] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] current.throw(*self._exc) [ 670.964822] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 670.964822] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] result = function(*args, **kwargs) [ 670.965164] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 670.965164] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] return func(*args, **kwargs) [ 670.965164] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 670.965164] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] raise e [ 670.965164] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 670.965164] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] nwinfo = self.network_api.allocate_for_instance( [ 670.965164] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 670.965164] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] created_port_ids = self._update_ports_for_instance( [ 670.965164] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 670.965164] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] with excutils.save_and_reraise_exception(): [ 670.965164] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 670.965164] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] self.force_reraise() [ 670.965164] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 670.965481] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] raise self.value [ 670.965481] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 670.965481] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] updated_port = self._update_port( [ 670.965481] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 670.965481] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] _ensure_no_port_binding_failure(port) [ 670.965481] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 670.965481] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] raise exception.PortBindingFailed(port_id=port['id']) [ 670.965481] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] nova.exception.PortBindingFailed: Binding failed for port 759c443f-21d6-4fef-a27b-6fe32cc20a63, please check neutron logs for more information. [ 670.965481] env[62460]: ERROR nova.compute.manager [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] [ 670.965481] env[62460]: DEBUG nova.compute.utils [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] Binding failed for port 759c443f-21d6-4fef-a27b-6fe32cc20a63, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 670.966513] env[62460]: DEBUG oslo_concurrency.lockutils [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.692s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 670.970063] env[62460]: DEBUG nova.compute.manager [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] Build of instance 01a3e1be-d29b-45d7-987e-66a4395ae2a4 was re-scheduled: Binding failed for port 759c443f-21d6-4fef-a27b-6fe32cc20a63, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 670.970063] env[62460]: DEBUG nova.compute.manager [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 670.970339] env[62460]: DEBUG oslo_concurrency.lockutils [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Acquiring lock "refresh_cache-01a3e1be-d29b-45d7-987e-66a4395ae2a4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 670.970534] env[62460]: DEBUG oslo_concurrency.lockutils [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Acquired lock "refresh_cache-01a3e1be-d29b-45d7-987e-66a4395ae2a4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 670.971208] env[62460]: DEBUG nova.network.neutron [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 671.135343] env[62460]: INFO nova.compute.manager [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] [instance: dba18c87-50a2-4ac8-8bd2-77f8f62b723c] Took 1.02 seconds to deallocate network for instance. [ 671.290576] env[62460]: DEBUG nova.compute.manager [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] [instance: 724a8083-2bde-483e-bd5a-a928def284bf] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 671.495461] env[62460]: DEBUG nova.network.neutron [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 671.646580] env[62460]: DEBUG nova.network.neutron [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.815692] env[62460]: DEBUG oslo_concurrency.lockutils [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 671.823498] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd283182-8039-4663-8be5-356b256c90da {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.831421] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a8a79ae-c13f-4784-b30f-5c1dd4c4aa85 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.861859] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1039207d-0d74-4462-9fa4-f7e0098bacab {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.869628] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfc39aa2-6df9-4995-8154-b439611c0d5f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.884584] env[62460]: DEBUG nova.compute.provider_tree [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 672.154350] env[62460]: DEBUG oslo_concurrency.lockutils [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Releasing lock "refresh_cache-01a3e1be-d29b-45d7-987e-66a4395ae2a4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 672.154577] env[62460]: DEBUG nova.compute.manager [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 672.154781] env[62460]: DEBUG nova.compute.manager [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 672.154949] env[62460]: DEBUG nova.network.neutron [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 672.189125] env[62460]: INFO nova.scheduler.client.report [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Deleted allocations for instance dba18c87-50a2-4ac8-8bd2-77f8f62b723c [ 672.193941] env[62460]: DEBUG nova.network.neutron [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 672.390194] env[62460]: DEBUG nova.scheduler.client.report [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 672.698301] env[62460]: DEBUG nova.network.neutron [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.699989] env[62460]: DEBUG oslo_concurrency.lockutils [None req-cfbab20e-8e8c-4877-9e84-39d632b2768d tempest-MigrationsAdminTest-857886804 tempest-MigrationsAdminTest-857886804-project-member] Lock "dba18c87-50a2-4ac8-8bd2-77f8f62b723c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 103.442s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 672.896983] env[62460]: DEBUG oslo_concurrency.lockutils [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.930s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 672.898039] env[62460]: ERROR nova.compute.manager [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4fa12c2e-8d85-4ab5-bec3-f905bdec71d5, please check neutron logs for more information. [ 672.898039] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] Traceback (most recent call last): [ 672.898039] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 672.898039] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] self.driver.spawn(context, instance, image_meta, [ 672.898039] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 672.898039] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 672.898039] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 672.898039] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] vm_ref = self.build_virtual_machine(instance, [ 672.898039] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 672.898039] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] vif_infos = vmwarevif.get_vif_info(self._session, [ 672.898039] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 672.898432] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] for vif in network_info: [ 672.898432] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 672.898432] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] return self._sync_wrapper(fn, *args, **kwargs) [ 672.898432] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 672.898432] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] self.wait() [ 672.898432] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 672.898432] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] self[:] = self._gt.wait() [ 672.898432] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 672.898432] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] return self._exit_event.wait() [ 672.898432] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 672.898432] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] current.throw(*self._exc) [ 672.898432] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 672.898432] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] result = function(*args, **kwargs) [ 672.898810] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 672.898810] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] return func(*args, **kwargs) [ 672.898810] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 672.898810] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] raise e [ 672.898810] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 672.898810] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] nwinfo = self.network_api.allocate_for_instance( [ 672.898810] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 672.898810] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] created_port_ids = self._update_ports_for_instance( [ 672.898810] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 672.898810] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] with excutils.save_and_reraise_exception(): [ 672.898810] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 672.898810] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] self.force_reraise() [ 672.898810] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 672.899215] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] raise self.value [ 672.899215] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 672.899215] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] updated_port = self._update_port( [ 672.899215] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 672.899215] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] _ensure_no_port_binding_failure(port) [ 672.899215] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 672.899215] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] raise exception.PortBindingFailed(port_id=port['id']) [ 672.899215] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] nova.exception.PortBindingFailed: Binding failed for port 4fa12c2e-8d85-4ab5-bec3-f905bdec71d5, please check neutron logs for more information. [ 672.899215] env[62460]: ERROR nova.compute.manager [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] [ 672.899215] env[62460]: DEBUG nova.compute.utils [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] Binding failed for port 4fa12c2e-8d85-4ab5-bec3-f905bdec71d5, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 672.902706] env[62460]: DEBUG oslo_concurrency.lockutils [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.186s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 672.906730] env[62460]: DEBUG nova.compute.manager [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] Build of instance 8c64f76d-cb18-41c5-9afa-e88af038f2b0 was re-scheduled: Binding failed for port 4fa12c2e-8d85-4ab5-bec3-f905bdec71d5, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 672.906844] env[62460]: DEBUG nova.compute.manager [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 672.908012] env[62460]: DEBUG oslo_concurrency.lockutils [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Acquiring lock "refresh_cache-8c64f76d-cb18-41c5-9afa-e88af038f2b0" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 672.908012] env[62460]: DEBUG oslo_concurrency.lockutils [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Acquired lock "refresh_cache-8c64f76d-cb18-41c5-9afa-e88af038f2b0" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 672.908012] env[62460]: DEBUG nova.network.neutron [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 673.201811] env[62460]: INFO nova.compute.manager [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] [instance: 01a3e1be-d29b-45d7-987e-66a4395ae2a4] Took 1.05 seconds to deallocate network for instance. [ 673.206443] env[62460]: DEBUG nova.compute.manager [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 673.440260] env[62460]: DEBUG nova.network.neutron [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 673.560492] env[62460]: DEBUG nova.network.neutron [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 673.741975] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 673.833913] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] Acquiring lock "4f3c7bd0-314a-45d2-a75d-a22b29d67cfe" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 673.834178] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] Lock "4f3c7bd0-314a-45d2-a75d-a22b29d67cfe" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 673.838488] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-324dcc69-2182-4903-8a17-40018cf782dc {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.846636] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54e09032-7148-48ef-9d68-c65cbaa19dc9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.877860] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45de2bfb-30f2-40b0-9c64-fa200a35745a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.886914] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a1c7588-afd9-4f77-bc74-80045ebba21b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.906014] env[62460]: DEBUG nova.compute.provider_tree [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 674.063680] env[62460]: DEBUG oslo_concurrency.lockutils [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Releasing lock "refresh_cache-8c64f76d-cb18-41c5-9afa-e88af038f2b0" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 674.063910] env[62460]: DEBUG nova.compute.manager [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 674.064112] env[62460]: DEBUG nova.compute.manager [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 674.064280] env[62460]: DEBUG nova.network.neutron [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 674.086346] env[62460]: DEBUG nova.network.neutron [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 674.251734] env[62460]: INFO nova.scheduler.client.report [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Deleted allocations for instance 01a3e1be-d29b-45d7-987e-66a4395ae2a4 [ 674.410163] env[62460]: DEBUG nova.scheduler.client.report [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 674.591235] env[62460]: DEBUG nova.network.neutron [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 674.770546] env[62460]: DEBUG oslo_concurrency.lockutils [None req-07247c91-c305-4886-ae01-0b0050f4571e tempest-SecurityGroupsTestJSON-1369561840 tempest-SecurityGroupsTestJSON-1369561840-project-member] Lock "01a3e1be-d29b-45d7-987e-66a4395ae2a4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 101.239s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 674.915785] env[62460]: DEBUG oslo_concurrency.lockutils [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.013s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 674.916443] env[62460]: ERROR nova.compute.manager [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 07d0ef21-dd97-42e1-8b07-67b1feade067, please check neutron logs for more information. [ 674.916443] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] Traceback (most recent call last): [ 674.916443] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 674.916443] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] self.driver.spawn(context, instance, image_meta, [ 674.916443] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 674.916443] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 674.916443] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 674.916443] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] vm_ref = self.build_virtual_machine(instance, [ 674.916443] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 674.916443] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] vif_infos = vmwarevif.get_vif_info(self._session, [ 674.916443] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 674.916760] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] for vif in network_info: [ 674.916760] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 674.916760] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] return self._sync_wrapper(fn, *args, **kwargs) [ 674.916760] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 674.916760] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] self.wait() [ 674.916760] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 674.916760] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] self[:] = self._gt.wait() [ 674.916760] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 674.916760] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] return self._exit_event.wait() [ 674.916760] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 674.916760] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] current.throw(*self._exc) [ 674.916760] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 674.916760] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] result = function(*args, **kwargs) [ 674.917082] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 674.917082] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] return func(*args, **kwargs) [ 674.917082] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 674.917082] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] raise e [ 674.917082] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 674.917082] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] nwinfo = self.network_api.allocate_for_instance( [ 674.917082] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 674.917082] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] created_port_ids = self._update_ports_for_instance( [ 674.917082] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 674.917082] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] with excutils.save_and_reraise_exception(): [ 674.917082] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 674.917082] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] self.force_reraise() [ 674.917082] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 674.917409] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] raise self.value [ 674.917409] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 674.917409] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] updated_port = self._update_port( [ 674.917409] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 674.917409] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] _ensure_no_port_binding_failure(port) [ 674.917409] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 674.917409] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] raise exception.PortBindingFailed(port_id=port['id']) [ 674.917409] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] nova.exception.PortBindingFailed: Binding failed for port 07d0ef21-dd97-42e1-8b07-67b1feade067, please check neutron logs for more information. [ 674.917409] env[62460]: ERROR nova.compute.manager [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] [ 674.917409] env[62460]: DEBUG nova.compute.utils [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] Binding failed for port 07d0ef21-dd97-42e1-8b07-67b1feade067, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 674.918578] env[62460]: DEBUG oslo_concurrency.lockutils [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.785s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 674.921527] env[62460]: DEBUG nova.compute.manager [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] Build of instance a24776d9-9950-4c83-9641-9675cabd5fd3 was re-scheduled: Binding failed for port 07d0ef21-dd97-42e1-8b07-67b1feade067, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 674.922098] env[62460]: DEBUG nova.compute.manager [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 674.922416] env[62460]: DEBUG oslo_concurrency.lockutils [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Acquiring lock "refresh_cache-a24776d9-9950-4c83-9641-9675cabd5fd3" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 674.922585] env[62460]: DEBUG oslo_concurrency.lockutils [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Acquired lock "refresh_cache-a24776d9-9950-4c83-9641-9675cabd5fd3" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 674.922836] env[62460]: DEBUG nova.network.neutron [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 675.096289] env[62460]: INFO nova.compute.manager [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 8c64f76d-cb18-41c5-9afa-e88af038f2b0] Took 1.03 seconds to deallocate network for instance. [ 675.273303] env[62460]: DEBUG nova.compute.manager [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 675.470764] env[62460]: DEBUG nova.network.neutron [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 675.602995] env[62460]: DEBUG nova.network.neutron [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 675.793302] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 675.802069] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50ef7185-0bcf-4774-ab43-cfe4f5dde638 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.808885] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8de9b008-c262-46ad-897e-d6edd50f2bd6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.839252] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90685ccd-b7b3-4e9d-a8c1-d6ac1a4db632 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.847111] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da826c0f-b778-45f1-99f6-7de95e969226 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.866427] env[62460]: DEBUG nova.compute.provider_tree [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 676.108654] env[62460]: DEBUG oslo_concurrency.lockutils [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Releasing lock "refresh_cache-a24776d9-9950-4c83-9641-9675cabd5fd3" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 676.108950] env[62460]: DEBUG nova.compute.manager [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 676.109151] env[62460]: DEBUG nova.compute.manager [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 676.109396] env[62460]: DEBUG nova.network.neutron [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 676.136233] env[62460]: DEBUG nova.network.neutron [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 676.142083] env[62460]: INFO nova.scheduler.client.report [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Deleted allocations for instance 8c64f76d-cb18-41c5-9afa-e88af038f2b0 [ 676.368950] env[62460]: DEBUG nova.scheduler.client.report [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 676.639331] env[62460]: DEBUG nova.network.neutron [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.652519] env[62460]: DEBUG oslo_concurrency.lockutils [None req-281d21cd-81da-48f9-84e1-4e72917e5218 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Lock "8c64f76d-cb18-41c5-9afa-e88af038f2b0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 97.982s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 676.838301] env[62460]: DEBUG oslo_concurrency.lockutils [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] Acquiring lock "559757ef-ee92-4b88-8631-6f743fb88bc6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 676.838556] env[62460]: DEBUG oslo_concurrency.lockutils [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] Lock "559757ef-ee92-4b88-8631-6f743fb88bc6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 676.875238] env[62460]: DEBUG oslo_concurrency.lockutils [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.957s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 676.875976] env[62460]: ERROR nova.compute.manager [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] [instance: 61426715-7a38-475d-895a-0eb6d6040c66] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8ecec905-4891-4a60-a063-ef5f68a1373b, please check neutron logs for more information. [ 676.875976] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] Traceback (most recent call last): [ 676.875976] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 676.875976] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] self.driver.spawn(context, instance, image_meta, [ 676.875976] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 676.875976] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] self._vmops.spawn(context, instance, image_meta, injected_files, [ 676.875976] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 676.875976] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] vm_ref = self.build_virtual_machine(instance, [ 676.875976] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 676.875976] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] vif_infos = vmwarevif.get_vif_info(self._session, [ 676.875976] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 676.876408] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] for vif in network_info: [ 676.876408] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 676.876408] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] return self._sync_wrapper(fn, *args, **kwargs) [ 676.876408] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 676.876408] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] self.wait() [ 676.876408] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 676.876408] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] self[:] = self._gt.wait() [ 676.876408] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 676.876408] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] return self._exit_event.wait() [ 676.876408] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 676.876408] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] current.throw(*self._exc) [ 676.876408] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 676.876408] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] result = function(*args, **kwargs) [ 676.876909] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 676.876909] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] return func(*args, **kwargs) [ 676.876909] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 676.876909] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] raise e [ 676.876909] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 676.876909] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] nwinfo = self.network_api.allocate_for_instance( [ 676.876909] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 676.876909] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] created_port_ids = self._update_ports_for_instance( [ 676.876909] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 676.876909] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] with excutils.save_and_reraise_exception(): [ 676.876909] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 676.876909] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] self.force_reraise() [ 676.876909] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 676.877243] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] raise self.value [ 676.877243] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 676.877243] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] updated_port = self._update_port( [ 676.877243] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 676.877243] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] _ensure_no_port_binding_failure(port) [ 676.877243] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 676.877243] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] raise exception.PortBindingFailed(port_id=port['id']) [ 676.877243] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] nova.exception.PortBindingFailed: Binding failed for port 8ecec905-4891-4a60-a063-ef5f68a1373b, please check neutron logs for more information. [ 676.877243] env[62460]: ERROR nova.compute.manager [instance: 61426715-7a38-475d-895a-0eb6d6040c66] [ 676.877243] env[62460]: DEBUG nova.compute.utils [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] [instance: 61426715-7a38-475d-895a-0eb6d6040c66] Binding failed for port 8ecec905-4891-4a60-a063-ef5f68a1373b, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 676.879865] env[62460]: DEBUG oslo_concurrency.lockutils [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.325s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 676.880413] env[62460]: INFO nova.compute.claims [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 676.883191] env[62460]: DEBUG nova.compute.manager [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] [instance: 61426715-7a38-475d-895a-0eb6d6040c66] Build of instance 61426715-7a38-475d-895a-0eb6d6040c66 was re-scheduled: Binding failed for port 8ecec905-4891-4a60-a063-ef5f68a1373b, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 676.883704] env[62460]: DEBUG nova.compute.manager [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] [instance: 61426715-7a38-475d-895a-0eb6d6040c66] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 676.883969] env[62460]: DEBUG oslo_concurrency.lockutils [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] Acquiring lock "refresh_cache-61426715-7a38-475d-895a-0eb6d6040c66" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 676.884172] env[62460]: DEBUG oslo_concurrency.lockutils [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] Acquired lock "refresh_cache-61426715-7a38-475d-895a-0eb6d6040c66" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 676.884363] env[62460]: DEBUG nova.network.neutron [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] [instance: 61426715-7a38-475d-895a-0eb6d6040c66] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 677.144131] env[62460]: INFO nova.compute.manager [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: a24776d9-9950-4c83-9641-9675cabd5fd3] Took 1.03 seconds to deallocate network for instance. [ 677.155400] env[62460]: DEBUG nova.compute.manager [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 677.431321] env[62460]: DEBUG nova.network.neutron [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] [instance: 61426715-7a38-475d-895a-0eb6d6040c66] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 677.670066] env[62460]: DEBUG nova.network.neutron [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] [instance: 61426715-7a38-475d-895a-0eb6d6040c66] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.685301] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 678.180754] env[62460]: DEBUG oslo_concurrency.lockutils [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] Releasing lock "refresh_cache-61426715-7a38-475d-895a-0eb6d6040c66" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 678.180754] env[62460]: DEBUG nova.compute.manager [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 678.180754] env[62460]: DEBUG nova.compute.manager [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] [instance: 61426715-7a38-475d-895a-0eb6d6040c66] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 678.180754] env[62460]: DEBUG nova.network.neutron [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] [instance: 61426715-7a38-475d-895a-0eb6d6040c66] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 678.203470] env[62460]: INFO nova.scheduler.client.report [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Deleted allocations for instance a24776d9-9950-4c83-9641-9675cabd5fd3 [ 678.210989] env[62460]: DEBUG nova.network.neutron [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] [instance: 61426715-7a38-475d-895a-0eb6d6040c66] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 678.267951] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18ed8d40-8774-4a81-a2ec-ec0d0429b4d9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.275983] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d84137c-21c6-4353-b0ae-f7f7bfd393c8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.306559] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5af0c9f2-da7d-428e-85ea-f215906b5040 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.313853] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-774bc122-bf0a-42e9-80d8-e347c97b413b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.330685] env[62460]: DEBUG nova.compute.provider_tree [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 678.672736] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Acquiring lock "588e5489-2f61-473a-90e9-883eb8b163c8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 678.673048] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Lock "588e5489-2f61-473a-90e9-883eb8b163c8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 678.714153] env[62460]: DEBUG nova.network.neutron [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] [instance: 61426715-7a38-475d-895a-0eb6d6040c66] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.716734] env[62460]: DEBUG oslo_concurrency.lockutils [None req-eb0a3026-d798-4d9d-a776-28043c6645a1 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Lock "a24776d9-9950-4c83-9641-9675cabd5fd3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 98.877s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 678.835018] env[62460]: DEBUG nova.scheduler.client.report [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 679.218676] env[62460]: INFO nova.compute.manager [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] [instance: 61426715-7a38-475d-895a-0eb6d6040c66] Took 1.04 seconds to deallocate network for instance. [ 679.225696] env[62460]: DEBUG nova.compute.manager [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 679.340194] env[62460]: DEBUG oslo_concurrency.lockutils [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.462s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 679.340735] env[62460]: DEBUG nova.compute.manager [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 679.344291] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.910s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 679.345567] env[62460]: INFO nova.compute.claims [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 679.594356] env[62460]: DEBUG oslo_concurrency.lockutils [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] Acquiring lock "6c58f9fe-fb6d-4012-8b69-39e4134996b2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 679.594356] env[62460]: DEBUG oslo_concurrency.lockutils [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] Lock "6c58f9fe-fb6d-4012-8b69-39e4134996b2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 679.754930] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 679.849954] env[62460]: DEBUG nova.compute.utils [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 679.853309] env[62460]: DEBUG nova.compute.manager [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 679.853487] env[62460]: DEBUG nova.network.neutron [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 679.908669] env[62460]: DEBUG nova.policy [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f602d798fe6b4c87a9496678069e6428', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c2e61d3de489431b9779f61231ebb46f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 680.272201] env[62460]: INFO nova.scheduler.client.report [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] Deleted allocations for instance 61426715-7a38-475d-895a-0eb6d6040c66 [ 680.357021] env[62460]: DEBUG nova.compute.manager [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 680.381683] env[62460]: DEBUG nova.network.neutron [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] Successfully created port: 1f0b1f33-f610-4a68-a5b9-d339068ba1bc {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 680.712149] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a0ced66-ba49-4cec-8c13-181451044c6a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.720480] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23babfe8-69e2-40e6-b17d-5b2c95655b1b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.750346] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3796150-d6ff-434f-b07e-626fa64d0282 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.757664] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe8df149-651a-42af-a085-3b98defd9aef {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.771482] env[62460]: DEBUG nova.compute.provider_tree [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 680.785273] env[62460]: DEBUG oslo_concurrency.lockutils [None req-71d90b6c-7608-4983-b112-dac02724ddb9 tempest-ServerPasswordTestJSON-1291359825 tempest-ServerPasswordTestJSON-1291359825-project-member] Lock "61426715-7a38-475d-895a-0eb6d6040c66" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 97.508s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 681.251885] env[62460]: DEBUG nova.compute.manager [req-8c8bcae3-63b7-48cb-9fab-d836f56a6f38 req-44f1cf02-6581-4d95-bba1-ddbaa1fcb320 service nova] [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] Received event network-changed-1f0b1f33-f610-4a68-a5b9-d339068ba1bc {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 681.252554] env[62460]: DEBUG nova.compute.manager [req-8c8bcae3-63b7-48cb-9fab-d836f56a6f38 req-44f1cf02-6581-4d95-bba1-ddbaa1fcb320 service nova] [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] Refreshing instance network info cache due to event network-changed-1f0b1f33-f610-4a68-a5b9-d339068ba1bc. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 681.253089] env[62460]: DEBUG oslo_concurrency.lockutils [req-8c8bcae3-63b7-48cb-9fab-d836f56a6f38 req-44f1cf02-6581-4d95-bba1-ddbaa1fcb320 service nova] Acquiring lock "refresh_cache-252df6ad-e29d-4596-bc99-4aae144bbcc5" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 681.253448] env[62460]: DEBUG oslo_concurrency.lockutils [req-8c8bcae3-63b7-48cb-9fab-d836f56a6f38 req-44f1cf02-6581-4d95-bba1-ddbaa1fcb320 service nova] Acquired lock "refresh_cache-252df6ad-e29d-4596-bc99-4aae144bbcc5" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 681.254035] env[62460]: DEBUG nova.network.neutron [req-8c8bcae3-63b7-48cb-9fab-d836f56a6f38 req-44f1cf02-6581-4d95-bba1-ddbaa1fcb320 service nova] [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] Refreshing network info cache for port 1f0b1f33-f610-4a68-a5b9-d339068ba1bc {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 681.277024] env[62460]: DEBUG nova.scheduler.client.report [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 681.287197] env[62460]: DEBUG nova.compute.manager [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 681.368569] env[62460]: DEBUG nova.compute.manager [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 681.392529] env[62460]: DEBUG nova.virt.hardware [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 681.392779] env[62460]: DEBUG nova.virt.hardware [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 681.392987] env[62460]: DEBUG nova.virt.hardware [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 681.393368] env[62460]: DEBUG nova.virt.hardware [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 681.393475] env[62460]: DEBUG nova.virt.hardware [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 681.393629] env[62460]: DEBUG nova.virt.hardware [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 681.393842] env[62460]: DEBUG nova.virt.hardware [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 681.394167] env[62460]: DEBUG nova.virt.hardware [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 681.394530] env[62460]: DEBUG nova.virt.hardware [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 681.394712] env[62460]: DEBUG nova.virt.hardware [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 681.394895] env[62460]: DEBUG nova.virt.hardware [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 681.397113] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-254ac2df-30c5-4410-bb05-f23523084e6e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.409364] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49985b45-2d20-4955-9365-25b7c94bddb1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.491305] env[62460]: ERROR nova.compute.manager [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1f0b1f33-f610-4a68-a5b9-d339068ba1bc, please check neutron logs for more information. [ 681.491305] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 681.491305] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 681.491305] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 681.491305] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 681.491305] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 681.491305] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 681.491305] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 681.491305] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 681.491305] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 681.491305] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 681.491305] env[62460]: ERROR nova.compute.manager raise self.value [ 681.491305] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 681.491305] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 681.491305] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 681.491305] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 681.494472] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 681.494472] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 681.494472] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1f0b1f33-f610-4a68-a5b9-d339068ba1bc, please check neutron logs for more information. [ 681.494472] env[62460]: ERROR nova.compute.manager [ 681.494472] env[62460]: Traceback (most recent call last): [ 681.494472] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 681.494472] env[62460]: listener.cb(fileno) [ 681.494472] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 681.494472] env[62460]: result = function(*args, **kwargs) [ 681.494472] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 681.494472] env[62460]: return func(*args, **kwargs) [ 681.494472] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 681.494472] env[62460]: raise e [ 681.494472] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 681.494472] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 681.494472] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 681.494472] env[62460]: created_port_ids = self._update_ports_for_instance( [ 681.494472] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 681.494472] env[62460]: with excutils.save_and_reraise_exception(): [ 681.494472] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 681.494472] env[62460]: self.force_reraise() [ 681.494472] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 681.494472] env[62460]: raise self.value [ 681.494472] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 681.494472] env[62460]: updated_port = self._update_port( [ 681.494472] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 681.494472] env[62460]: _ensure_no_port_binding_failure(port) [ 681.494472] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 681.494472] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 681.495216] env[62460]: nova.exception.PortBindingFailed: Binding failed for port 1f0b1f33-f610-4a68-a5b9-d339068ba1bc, please check neutron logs for more information. [ 681.495216] env[62460]: Removing descriptor: 19 [ 681.495216] env[62460]: ERROR nova.compute.manager [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1f0b1f33-f610-4a68-a5b9-d339068ba1bc, please check neutron logs for more information. [ 681.495216] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] Traceback (most recent call last): [ 681.495216] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 681.495216] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] yield resources [ 681.495216] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 681.495216] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] self.driver.spawn(context, instance, image_meta, [ 681.495216] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 681.495216] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 681.495216] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 681.495216] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] vm_ref = self.build_virtual_machine(instance, [ 681.495911] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 681.495911] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] vif_infos = vmwarevif.get_vif_info(self._session, [ 681.495911] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 681.495911] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] for vif in network_info: [ 681.495911] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 681.495911] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] return self._sync_wrapper(fn, *args, **kwargs) [ 681.495911] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 681.495911] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] self.wait() [ 681.495911] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 681.495911] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] self[:] = self._gt.wait() [ 681.495911] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 681.495911] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] return self._exit_event.wait() [ 681.495911] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 681.496263] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] result = hub.switch() [ 681.496263] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 681.496263] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] return self.greenlet.switch() [ 681.496263] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 681.496263] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] result = function(*args, **kwargs) [ 681.496263] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 681.496263] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] return func(*args, **kwargs) [ 681.496263] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 681.496263] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] raise e [ 681.496263] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 681.496263] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] nwinfo = self.network_api.allocate_for_instance( [ 681.496263] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 681.496263] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] created_port_ids = self._update_ports_for_instance( [ 681.496581] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 681.496581] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] with excutils.save_and_reraise_exception(): [ 681.496581] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 681.496581] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] self.force_reraise() [ 681.496581] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 681.496581] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] raise self.value [ 681.496581] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 681.496581] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] updated_port = self._update_port( [ 681.496581] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 681.496581] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] _ensure_no_port_binding_failure(port) [ 681.496581] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 681.496581] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] raise exception.PortBindingFailed(port_id=port['id']) [ 681.496872] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] nova.exception.PortBindingFailed: Binding failed for port 1f0b1f33-f610-4a68-a5b9-d339068ba1bc, please check neutron logs for more information. [ 681.496872] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] [ 681.496872] env[62460]: INFO nova.compute.manager [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] Terminating instance [ 681.498071] env[62460]: DEBUG oslo_concurrency.lockutils [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] Acquiring lock "refresh_cache-252df6ad-e29d-4596-bc99-4aae144bbcc5" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 681.776337] env[62460]: DEBUG nova.network.neutron [req-8c8bcae3-63b7-48cb-9fab-d836f56a6f38 req-44f1cf02-6581-4d95-bba1-ddbaa1fcb320 service nova] [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 681.780691] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.436s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 681.781082] env[62460]: DEBUG nova.compute.manager [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 681.784605] env[62460]: DEBUG oslo_concurrency.lockutils [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.592s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 681.814173] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 681.927234] env[62460]: DEBUG nova.network.neutron [req-8c8bcae3-63b7-48cb-9fab-d836f56a6f38 req-44f1cf02-6581-4d95-bba1-ddbaa1fcb320 service nova] [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.293491] env[62460]: DEBUG nova.compute.utils [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 682.297185] env[62460]: DEBUG nova.compute.manager [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 682.298433] env[62460]: DEBUG nova.network.neutron [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 682.304229] env[62460]: DEBUG oslo_concurrency.lockutils [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Acquiring lock "c7d51b39-b449-4af3-a4d0-c746983ded3e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 682.304798] env[62460]: DEBUG oslo_concurrency.lockutils [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Lock "c7d51b39-b449-4af3-a4d0-c746983ded3e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 682.429191] env[62460]: DEBUG oslo_concurrency.lockutils [req-8c8bcae3-63b7-48cb-9fab-d836f56a6f38 req-44f1cf02-6581-4d95-bba1-ddbaa1fcb320 service nova] Releasing lock "refresh_cache-252df6ad-e29d-4596-bc99-4aae144bbcc5" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 682.429593] env[62460]: DEBUG oslo_concurrency.lockutils [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] Acquired lock "refresh_cache-252df6ad-e29d-4596-bc99-4aae144bbcc5" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 682.429793] env[62460]: DEBUG nova.network.neutron [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 682.491242] env[62460]: DEBUG nova.policy [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7684644b29ad4d709309cadf8c154449', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '19e53e9f113e4465924f75c1f9a39008', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 682.635601] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0052f9dc-2a46-48a8-b791-952f75222dcc {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.644774] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6a71559-26fc-4dc6-8fc8-9011e5f72a02 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.685200] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e087deb7-737c-427d-888e-4f42a04e0dc3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.692947] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b68ab346-25ca-44e0-aeb1-9b310c475404 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.707464] env[62460]: DEBUG nova.compute.provider_tree [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 682.804504] env[62460]: DEBUG nova.compute.manager [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 682.959575] env[62460]: DEBUG nova.network.neutron [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 683.028721] env[62460]: DEBUG nova.network.neutron [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] Successfully created port: 956ced4a-8d22-4660-b90c-a7a197dd8adb {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 683.060166] env[62460]: DEBUG nova.network.neutron [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 683.211887] env[62460]: DEBUG nova.scheduler.client.report [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 683.276652] env[62460]: DEBUG nova.compute.manager [req-1e1911d7-945e-4436-83e3-ef6ffdef45ff req-ba54262b-0dd9-49d8-8d9d-18d6336db262 service nova] [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] Received event network-vif-deleted-1f0b1f33-f610-4a68-a5b9-d339068ba1bc {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 683.457393] env[62460]: DEBUG nova.network.neutron [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] Successfully created port: 970bf89a-a396-4881-9056-fdadbb9f0783 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 683.565133] env[62460]: DEBUG oslo_concurrency.lockutils [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] Releasing lock "refresh_cache-252df6ad-e29d-4596-bc99-4aae144bbcc5" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 683.565678] env[62460]: DEBUG nova.compute.manager [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 683.566244] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 683.567055] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2c7abc9c-eff3-4913-adc0-0d3889ba373b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.576410] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aad994e3-3bf9-459a-b8d8-fafdc12999e6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.598436] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 252df6ad-e29d-4596-bc99-4aae144bbcc5 could not be found. [ 683.598511] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 683.598631] env[62460]: INFO nova.compute.manager [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] Took 0.03 seconds to destroy the instance on the hypervisor. [ 683.598894] env[62460]: DEBUG oslo.service.loopingcall [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 683.599135] env[62460]: DEBUG nova.compute.manager [-] [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 683.599224] env[62460]: DEBUG nova.network.neutron [-] [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 683.620742] env[62460]: DEBUG nova.network.neutron [-] [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 683.715997] env[62460]: DEBUG oslo_concurrency.lockutils [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.932s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 683.716649] env[62460]: ERROR nova.compute.manager [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 58dfc745-c39d-4558-af1a-1f04af4a3b87, please check neutron logs for more information. [ 683.716649] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] Traceback (most recent call last): [ 683.716649] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 683.716649] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] self.driver.spawn(context, instance, image_meta, [ 683.716649] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 683.716649] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 683.716649] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 683.716649] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] vm_ref = self.build_virtual_machine(instance, [ 683.716649] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 683.716649] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] vif_infos = vmwarevif.get_vif_info(self._session, [ 683.716649] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 683.716980] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] for vif in network_info: [ 683.716980] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 683.716980] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] return self._sync_wrapper(fn, *args, **kwargs) [ 683.716980] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 683.716980] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] self.wait() [ 683.716980] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 683.716980] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] self[:] = self._gt.wait() [ 683.716980] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 683.716980] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] return self._exit_event.wait() [ 683.716980] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 683.716980] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] current.throw(*self._exc) [ 683.716980] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 683.716980] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] result = function(*args, **kwargs) [ 683.717862] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 683.717862] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] return func(*args, **kwargs) [ 683.717862] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 683.717862] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] raise e [ 683.717862] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 683.717862] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] nwinfo = self.network_api.allocate_for_instance( [ 683.717862] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 683.717862] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] created_port_ids = self._update_ports_for_instance( [ 683.717862] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 683.717862] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] with excutils.save_and_reraise_exception(): [ 683.717862] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 683.717862] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] self.force_reraise() [ 683.717862] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 683.718490] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] raise self.value [ 683.718490] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 683.718490] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] updated_port = self._update_port( [ 683.718490] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 683.718490] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] _ensure_no_port_binding_failure(port) [ 683.718490] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 683.718490] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] raise exception.PortBindingFailed(port_id=port['id']) [ 683.718490] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] nova.exception.PortBindingFailed: Binding failed for port 58dfc745-c39d-4558-af1a-1f04af4a3b87, please check neutron logs for more information. [ 683.718490] env[62460]: ERROR nova.compute.manager [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] [ 683.718490] env[62460]: DEBUG nova.compute.utils [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] Binding failed for port 58dfc745-c39d-4558-af1a-1f04af4a3b87, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 683.719735] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.837s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 683.721555] env[62460]: INFO nova.compute.claims [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] [instance: 578627be-e695-4953-8d0e-9763d12b9a28] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 683.728023] env[62460]: DEBUG nova.compute.manager [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] Build of instance b6efad50-aa2e-49f9-9ce7-5fead31db7a8 was re-scheduled: Binding failed for port 58dfc745-c39d-4558-af1a-1f04af4a3b87, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 683.728023] env[62460]: DEBUG nova.compute.manager [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 683.728023] env[62460]: DEBUG oslo_concurrency.lockutils [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "refresh_cache-b6efad50-aa2e-49f9-9ce7-5fead31db7a8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 683.728023] env[62460]: DEBUG oslo_concurrency.lockutils [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquired lock "refresh_cache-b6efad50-aa2e-49f9-9ce7-5fead31db7a8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 683.728306] env[62460]: DEBUG nova.network.neutron [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 683.827110] env[62460]: DEBUG nova.compute.manager [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 683.854401] env[62460]: DEBUG nova.virt.hardware [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 683.854647] env[62460]: DEBUG nova.virt.hardware [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 683.854807] env[62460]: DEBUG nova.virt.hardware [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 683.854987] env[62460]: DEBUG nova.virt.hardware [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 683.858560] env[62460]: DEBUG nova.virt.hardware [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 683.858560] env[62460]: DEBUG nova.virt.hardware [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 683.858560] env[62460]: DEBUG nova.virt.hardware [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 683.858560] env[62460]: DEBUG nova.virt.hardware [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 683.858560] env[62460]: DEBUG nova.virt.hardware [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 683.858708] env[62460]: DEBUG nova.virt.hardware [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 683.858708] env[62460]: DEBUG nova.virt.hardware [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 683.858708] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31833525-848a-4ad4-9be2-94c6e68c229c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.867671] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3d61dc7-6e19-4e11-9ebf-52af7bd513a2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.123563] env[62460]: DEBUG nova.network.neutron [-] [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.260167] env[62460]: DEBUG nova.network.neutron [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 684.578787] env[62460]: DEBUG nova.network.neutron [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.582556] env[62460]: ERROR nova.compute.manager [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 956ced4a-8d22-4660-b90c-a7a197dd8adb, please check neutron logs for more information. [ 684.582556] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 684.582556] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 684.582556] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 684.582556] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 684.582556] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 684.582556] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 684.582556] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 684.582556] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 684.582556] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 684.582556] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 684.582556] env[62460]: ERROR nova.compute.manager raise self.value [ 684.582556] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 684.582556] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 684.582556] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 684.582556] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 684.583132] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 684.583132] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 684.583132] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 956ced4a-8d22-4660-b90c-a7a197dd8adb, please check neutron logs for more information. [ 684.583132] env[62460]: ERROR nova.compute.manager [ 684.583132] env[62460]: Traceback (most recent call last): [ 684.583132] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 684.583132] env[62460]: listener.cb(fileno) [ 684.583132] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 684.583132] env[62460]: result = function(*args, **kwargs) [ 684.583132] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 684.583132] env[62460]: return func(*args, **kwargs) [ 684.583132] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 684.583132] env[62460]: raise e [ 684.583132] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 684.583132] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 684.583132] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 684.583132] env[62460]: created_port_ids = self._update_ports_for_instance( [ 684.583132] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 684.583132] env[62460]: with excutils.save_and_reraise_exception(): [ 684.583132] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 684.583132] env[62460]: self.force_reraise() [ 684.583132] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 684.583132] env[62460]: raise self.value [ 684.583132] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 684.583132] env[62460]: updated_port = self._update_port( [ 684.583132] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 684.583132] env[62460]: _ensure_no_port_binding_failure(port) [ 684.583132] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 684.583132] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 684.583981] env[62460]: nova.exception.PortBindingFailed: Binding failed for port 956ced4a-8d22-4660-b90c-a7a197dd8adb, please check neutron logs for more information. [ 684.583981] env[62460]: Removing descriptor: 19 [ 684.583981] env[62460]: ERROR nova.compute.manager [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 956ced4a-8d22-4660-b90c-a7a197dd8adb, please check neutron logs for more information. [ 684.583981] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] Traceback (most recent call last): [ 684.583981] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 684.583981] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] yield resources [ 684.583981] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 684.583981] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] self.driver.spawn(context, instance, image_meta, [ 684.583981] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 684.583981] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 684.583981] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 684.583981] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] vm_ref = self.build_virtual_machine(instance, [ 684.584369] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 684.584369] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] vif_infos = vmwarevif.get_vif_info(self._session, [ 684.584369] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 684.584369] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] for vif in network_info: [ 684.584369] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 684.584369] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] return self._sync_wrapper(fn, *args, **kwargs) [ 684.584369] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 684.584369] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] self.wait() [ 684.584369] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 684.584369] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] self[:] = self._gt.wait() [ 684.584369] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 684.584369] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] return self._exit_event.wait() [ 684.584369] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 684.584772] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] result = hub.switch() [ 684.584772] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 684.584772] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] return self.greenlet.switch() [ 684.584772] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 684.584772] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] result = function(*args, **kwargs) [ 684.584772] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 684.584772] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] return func(*args, **kwargs) [ 684.584772] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 684.584772] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] raise e [ 684.584772] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 684.584772] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] nwinfo = self.network_api.allocate_for_instance( [ 684.584772] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 684.584772] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] created_port_ids = self._update_ports_for_instance( [ 684.585218] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 684.585218] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] with excutils.save_and_reraise_exception(): [ 684.585218] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 684.585218] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] self.force_reraise() [ 684.585218] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 684.585218] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] raise self.value [ 684.585218] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 684.585218] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] updated_port = self._update_port( [ 684.585218] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 684.585218] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] _ensure_no_port_binding_failure(port) [ 684.585218] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 684.585218] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] raise exception.PortBindingFailed(port_id=port['id']) [ 684.585574] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] nova.exception.PortBindingFailed: Binding failed for port 956ced4a-8d22-4660-b90c-a7a197dd8adb, please check neutron logs for more information. [ 684.585574] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] [ 684.585574] env[62460]: INFO nova.compute.manager [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] Terminating instance [ 684.587307] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Acquiring lock "refresh_cache-02cb219f-989f-45b5-a9ba-d0b5a521cfa2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 684.587474] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Acquired lock "refresh_cache-02cb219f-989f-45b5-a9ba-d0b5a521cfa2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 684.591105] env[62460]: DEBUG nova.network.neutron [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 684.628836] env[62460]: INFO nova.compute.manager [-] [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] Took 1.03 seconds to deallocate network for instance. [ 684.631075] env[62460]: DEBUG nova.compute.claims [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 684.631274] env[62460]: DEBUG oslo_concurrency.lockutils [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 684.673134] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Acquiring lock "51718896-f5bf-43a9-9396-1ac768737ba2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 684.673134] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Lock "51718896-f5bf-43a9-9396-1ac768737ba2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 685.047202] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac0bf778-dad3-4bba-8709-26e44b8be629 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.055271] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1423905-57eb-4e8f-9f39-8b56f86c76cd {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.085989] env[62460]: DEBUG oslo_concurrency.lockutils [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Releasing lock "refresh_cache-b6efad50-aa2e-49f9-9ce7-5fead31db7a8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 685.086225] env[62460]: DEBUG nova.compute.manager [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 685.086279] env[62460]: DEBUG nova.compute.manager [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 685.086732] env[62460]: DEBUG nova.network.neutron [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 685.088988] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2849ff43-a3c2-4219-a6b6-b5a5f040ae4d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.102012] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2111d7a1-7431-44d9-9414-3fd196fc0161 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.113789] env[62460]: DEBUG nova.compute.provider_tree [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 685.115382] env[62460]: DEBUG nova.network.neutron [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 685.165668] env[62460]: DEBUG nova.network.neutron [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 685.288951] env[62460]: DEBUG nova.network.neutron [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 685.306933] env[62460]: DEBUG nova.compute.manager [req-72b9ac8f-3125-4bbe-8705-2239ea717ba9 req-088ca476-6b19-40d3-a01d-eb0aa0f30f5e service nova] [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] Received event network-changed-956ced4a-8d22-4660-b90c-a7a197dd8adb {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 685.307074] env[62460]: DEBUG nova.compute.manager [req-72b9ac8f-3125-4bbe-8705-2239ea717ba9 req-088ca476-6b19-40d3-a01d-eb0aa0f30f5e service nova] [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] Refreshing instance network info cache due to event network-changed-956ced4a-8d22-4660-b90c-a7a197dd8adb. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 685.307274] env[62460]: DEBUG oslo_concurrency.lockutils [req-72b9ac8f-3125-4bbe-8705-2239ea717ba9 req-088ca476-6b19-40d3-a01d-eb0aa0f30f5e service nova] Acquiring lock "refresh_cache-02cb219f-989f-45b5-a9ba-d0b5a521cfa2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 685.617771] env[62460]: DEBUG nova.scheduler.client.report [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 685.620984] env[62460]: DEBUG nova.network.neutron [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 685.685168] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Acquiring lock "1f318a64-2c38-470b-8fae-4ba4543a5681" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 685.685365] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Lock "1f318a64-2c38-470b-8fae-4ba4543a5681" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 685.792133] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Releasing lock "refresh_cache-02cb219f-989f-45b5-a9ba-d0b5a521cfa2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 685.792614] env[62460]: DEBUG nova.compute.manager [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 685.792837] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 685.793242] env[62460]: DEBUG oslo_concurrency.lockutils [req-72b9ac8f-3125-4bbe-8705-2239ea717ba9 req-088ca476-6b19-40d3-a01d-eb0aa0f30f5e service nova] Acquired lock "refresh_cache-02cb219f-989f-45b5-a9ba-d0b5a521cfa2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 685.793442] env[62460]: DEBUG nova.network.neutron [req-72b9ac8f-3125-4bbe-8705-2239ea717ba9 req-088ca476-6b19-40d3-a01d-eb0aa0f30f5e service nova] [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] Refreshing network info cache for port 956ced4a-8d22-4660-b90c-a7a197dd8adb {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 685.795719] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0c6ea080-a15f-48b8-985d-af3b89009882 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.806138] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b8d50dc-3119-40ca-b2cc-b2c038792836 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.832034] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 02cb219f-989f-45b5-a9ba-d0b5a521cfa2 could not be found. [ 685.832034] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 685.832034] env[62460]: INFO nova.compute.manager [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] Took 0.04 seconds to destroy the instance on the hypervisor. [ 685.832034] env[62460]: DEBUG oslo.service.loopingcall [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 685.832282] env[62460]: DEBUG nova.compute.manager [-] [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 685.832424] env[62460]: DEBUG nova.network.neutron [-] [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 685.865952] env[62460]: DEBUG nova.network.neutron [-] [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 686.124176] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.404s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 686.124791] env[62460]: DEBUG nova.compute.manager [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] [instance: 578627be-e695-4953-8d0e-9763d12b9a28] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 686.127591] env[62460]: INFO nova.compute.manager [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: b6efad50-aa2e-49f9-9ce7-5fead31db7a8] Took 1.04 seconds to deallocate network for instance. [ 686.130296] env[62460]: DEBUG oslo_concurrency.lockutils [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.315s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 686.133640] env[62460]: INFO nova.compute.claims [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] [instance: 724a8083-2bde-483e-bd5a-a928def284bf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 686.315088] env[62460]: DEBUG nova.network.neutron [req-72b9ac8f-3125-4bbe-8705-2239ea717ba9 req-088ca476-6b19-40d3-a01d-eb0aa0f30f5e service nova] [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 686.418829] env[62460]: DEBUG nova.network.neutron [req-72b9ac8f-3125-4bbe-8705-2239ea717ba9 req-088ca476-6b19-40d3-a01d-eb0aa0f30f5e service nova] [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 686.573933] env[62460]: DEBUG nova.network.neutron [-] [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 686.632232] env[62460]: DEBUG nova.compute.utils [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 686.633589] env[62460]: DEBUG nova.compute.manager [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] [instance: 578627be-e695-4953-8d0e-9763d12b9a28] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 686.633762] env[62460]: DEBUG nova.network.neutron [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] [instance: 578627be-e695-4953-8d0e-9763d12b9a28] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 686.673384] env[62460]: DEBUG nova.policy [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '33c6dd49d5554f8abb6711b3e84f164c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4e3a09db827a4fe096c2bf492917703a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 686.922644] env[62460]: DEBUG oslo_concurrency.lockutils [req-72b9ac8f-3125-4bbe-8705-2239ea717ba9 req-088ca476-6b19-40d3-a01d-eb0aa0f30f5e service nova] Releasing lock "refresh_cache-02cb219f-989f-45b5-a9ba-d0b5a521cfa2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 686.922644] env[62460]: DEBUG nova.compute.manager [req-72b9ac8f-3125-4bbe-8705-2239ea717ba9 req-088ca476-6b19-40d3-a01d-eb0aa0f30f5e service nova] [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] Received event network-vif-deleted-956ced4a-8d22-4660-b90c-a7a197dd8adb {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 686.971613] env[62460]: DEBUG nova.network.neutron [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] [instance: 578627be-e695-4953-8d0e-9763d12b9a28] Successfully created port: fa6e34fb-aa15-4b5a-909a-4ed21e7e6fbd {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 687.076693] env[62460]: INFO nova.compute.manager [-] [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] Took 1.24 seconds to deallocate network for instance. [ 687.079359] env[62460]: DEBUG nova.compute.claims [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 687.079634] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 687.137430] env[62460]: DEBUG nova.compute.manager [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] [instance: 578627be-e695-4953-8d0e-9763d12b9a28] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 687.177381] env[62460]: INFO nova.scheduler.client.report [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Deleted allocations for instance b6efad50-aa2e-49f9-9ce7-5fead31db7a8 [ 687.512381] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4038b504-8c75-417f-9505-f5e2e946e174 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.519675] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbed6791-76e7-4b2f-aafd-0db23ab54d6f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.549457] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a30f5973-1fee-43c1-a957-88eaf800f2ae {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.556465] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7041fa4f-a526-4fbe-ace2-7092e211a070 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.571054] env[62460]: DEBUG nova.compute.provider_tree [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 687.687574] env[62460]: DEBUG oslo_concurrency.lockutils [None req-961f3dde-de3a-4a3c-849e-d1e9c814e2d7 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "b6efad50-aa2e-49f9-9ce7-5fead31db7a8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 101.490s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 687.740153] env[62460]: DEBUG nova.compute.manager [req-24c25ac7-219d-43ed-9cba-d223cfffee03 req-78bacdc4-f177-4db4-80c3-091e51ca5792 service nova] [instance: 578627be-e695-4953-8d0e-9763d12b9a28] Received event network-changed-fa6e34fb-aa15-4b5a-909a-4ed21e7e6fbd {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 687.740355] env[62460]: DEBUG nova.compute.manager [req-24c25ac7-219d-43ed-9cba-d223cfffee03 req-78bacdc4-f177-4db4-80c3-091e51ca5792 service nova] [instance: 578627be-e695-4953-8d0e-9763d12b9a28] Refreshing instance network info cache due to event network-changed-fa6e34fb-aa15-4b5a-909a-4ed21e7e6fbd. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 687.740567] env[62460]: DEBUG oslo_concurrency.lockutils [req-24c25ac7-219d-43ed-9cba-d223cfffee03 req-78bacdc4-f177-4db4-80c3-091e51ca5792 service nova] Acquiring lock "refresh_cache-578627be-e695-4953-8d0e-9763d12b9a28" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 687.740710] env[62460]: DEBUG oslo_concurrency.lockutils [req-24c25ac7-219d-43ed-9cba-d223cfffee03 req-78bacdc4-f177-4db4-80c3-091e51ca5792 service nova] Acquired lock "refresh_cache-578627be-e695-4953-8d0e-9763d12b9a28" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 687.740946] env[62460]: DEBUG nova.network.neutron [req-24c25ac7-219d-43ed-9cba-d223cfffee03 req-78bacdc4-f177-4db4-80c3-091e51ca5792 service nova] [instance: 578627be-e695-4953-8d0e-9763d12b9a28] Refreshing network info cache for port fa6e34fb-aa15-4b5a-909a-4ed21e7e6fbd {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 688.014041] env[62460]: ERROR nova.compute.manager [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port fa6e34fb-aa15-4b5a-909a-4ed21e7e6fbd, please check neutron logs for more information. [ 688.014041] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 688.014041] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 688.014041] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 688.014041] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 688.014041] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 688.014041] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 688.014041] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 688.014041] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 688.014041] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 688.014041] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 688.014041] env[62460]: ERROR nova.compute.manager raise self.value [ 688.014041] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 688.014041] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 688.014041] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 688.014041] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 688.014577] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 688.014577] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 688.014577] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port fa6e34fb-aa15-4b5a-909a-4ed21e7e6fbd, please check neutron logs for more information. [ 688.014577] env[62460]: ERROR nova.compute.manager [ 688.014577] env[62460]: Traceback (most recent call last): [ 688.014577] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 688.014577] env[62460]: listener.cb(fileno) [ 688.014577] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 688.014577] env[62460]: result = function(*args, **kwargs) [ 688.014577] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 688.014577] env[62460]: return func(*args, **kwargs) [ 688.014577] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 688.014577] env[62460]: raise e [ 688.014577] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 688.014577] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 688.014577] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 688.014577] env[62460]: created_port_ids = self._update_ports_for_instance( [ 688.014577] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 688.014577] env[62460]: with excutils.save_and_reraise_exception(): [ 688.014577] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 688.014577] env[62460]: self.force_reraise() [ 688.014577] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 688.014577] env[62460]: raise self.value [ 688.014577] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 688.014577] env[62460]: updated_port = self._update_port( [ 688.014577] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 688.014577] env[62460]: _ensure_no_port_binding_failure(port) [ 688.014577] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 688.014577] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 688.015313] env[62460]: nova.exception.PortBindingFailed: Binding failed for port fa6e34fb-aa15-4b5a-909a-4ed21e7e6fbd, please check neutron logs for more information. [ 688.015313] env[62460]: Removing descriptor: 18 [ 688.074705] env[62460]: DEBUG nova.scheduler.client.report [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 688.147418] env[62460]: DEBUG nova.compute.manager [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] [instance: 578627be-e695-4953-8d0e-9763d12b9a28] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 688.175826] env[62460]: DEBUG nova.virt.hardware [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 688.176106] env[62460]: DEBUG nova.virt.hardware [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 688.176269] env[62460]: DEBUG nova.virt.hardware [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 688.176451] env[62460]: DEBUG nova.virt.hardware [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 688.176596] env[62460]: DEBUG nova.virt.hardware [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 688.176741] env[62460]: DEBUG nova.virt.hardware [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 688.176945] env[62460]: DEBUG nova.virt.hardware [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 688.177134] env[62460]: DEBUG nova.virt.hardware [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 688.177309] env[62460]: DEBUG nova.virt.hardware [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 688.177473] env[62460]: DEBUG nova.virt.hardware [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 688.177646] env[62460]: DEBUG nova.virt.hardware [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 688.178532] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2b8f9dc-f004-410c-9fd0-e0a3b12bf417 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.186410] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7364b03-5140-4f0b-a507-e9e1bc9a37cb {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.190471] env[62460]: DEBUG nova.compute.manager [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 688.202429] env[62460]: ERROR nova.compute.manager [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] [instance: 578627be-e695-4953-8d0e-9763d12b9a28] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port fa6e34fb-aa15-4b5a-909a-4ed21e7e6fbd, please check neutron logs for more information. [ 688.202429] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] Traceback (most recent call last): [ 688.202429] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 688.202429] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] yield resources [ 688.202429] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 688.202429] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] self.driver.spawn(context, instance, image_meta, [ 688.202429] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 688.202429] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] self._vmops.spawn(context, instance, image_meta, injected_files, [ 688.202429] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 688.202429] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] vm_ref = self.build_virtual_machine(instance, [ 688.202429] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 688.202836] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] vif_infos = vmwarevif.get_vif_info(self._session, [ 688.202836] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 688.202836] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] for vif in network_info: [ 688.202836] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 688.202836] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] return self._sync_wrapper(fn, *args, **kwargs) [ 688.202836] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 688.202836] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] self.wait() [ 688.202836] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 688.202836] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] self[:] = self._gt.wait() [ 688.202836] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 688.202836] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] return self._exit_event.wait() [ 688.202836] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 688.202836] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] current.throw(*self._exc) [ 688.203247] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 688.203247] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] result = function(*args, **kwargs) [ 688.203247] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 688.203247] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] return func(*args, **kwargs) [ 688.203247] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 688.203247] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] raise e [ 688.203247] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 688.203247] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] nwinfo = self.network_api.allocate_for_instance( [ 688.203247] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 688.203247] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] created_port_ids = self._update_ports_for_instance( [ 688.203247] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 688.203247] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] with excutils.save_and_reraise_exception(): [ 688.203247] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 688.203653] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] self.force_reraise() [ 688.203653] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 688.203653] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] raise self.value [ 688.203653] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 688.203653] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] updated_port = self._update_port( [ 688.203653] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 688.203653] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] _ensure_no_port_binding_failure(port) [ 688.203653] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 688.203653] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] raise exception.PortBindingFailed(port_id=port['id']) [ 688.203653] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] nova.exception.PortBindingFailed: Binding failed for port fa6e34fb-aa15-4b5a-909a-4ed21e7e6fbd, please check neutron logs for more information. [ 688.203653] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] [ 688.203653] env[62460]: INFO nova.compute.manager [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] [instance: 578627be-e695-4953-8d0e-9763d12b9a28] Terminating instance [ 688.204588] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] Acquiring lock "refresh_cache-578627be-e695-4953-8d0e-9763d12b9a28" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 688.265266] env[62460]: DEBUG nova.network.neutron [req-24c25ac7-219d-43ed-9cba-d223cfffee03 req-78bacdc4-f177-4db4-80c3-091e51ca5792 service nova] [instance: 578627be-e695-4953-8d0e-9763d12b9a28] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 688.417415] env[62460]: DEBUG nova.network.neutron [req-24c25ac7-219d-43ed-9cba-d223cfffee03 req-78bacdc4-f177-4db4-80c3-091e51ca5792 service nova] [instance: 578627be-e695-4953-8d0e-9763d12b9a28] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 688.579480] env[62460]: DEBUG oslo_concurrency.lockutils [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.449s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 688.580109] env[62460]: DEBUG nova.compute.manager [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] [instance: 724a8083-2bde-483e-bd5a-a928def284bf] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 688.582810] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.841s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 688.584580] env[62460]: INFO nova.compute.claims [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 688.714907] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 688.919746] env[62460]: DEBUG oslo_concurrency.lockutils [req-24c25ac7-219d-43ed-9cba-d223cfffee03 req-78bacdc4-f177-4db4-80c3-091e51ca5792 service nova] Releasing lock "refresh_cache-578627be-e695-4953-8d0e-9763d12b9a28" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 688.920212] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] Acquired lock "refresh_cache-578627be-e695-4953-8d0e-9763d12b9a28" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 688.920407] env[62460]: DEBUG nova.network.neutron [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] [instance: 578627be-e695-4953-8d0e-9763d12b9a28] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 689.088857] env[62460]: DEBUG nova.compute.utils [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 689.093601] env[62460]: DEBUG nova.compute.manager [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] [instance: 724a8083-2bde-483e-bd5a-a928def284bf] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 689.093762] env[62460]: DEBUG nova.network.neutron [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] [instance: 724a8083-2bde-483e-bd5a-a928def284bf] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 689.146876] env[62460]: DEBUG nova.policy [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '39bb4ebddf1e49a69644ad6bbe019060', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd5717866e8b649f0b4e0e3dddc5a1424', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 689.448285] env[62460]: DEBUG nova.network.neutron [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] [instance: 578627be-e695-4953-8d0e-9763d12b9a28] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 689.479989] env[62460]: DEBUG nova.network.neutron [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] [instance: 724a8083-2bde-483e-bd5a-a928def284bf] Successfully created port: 7464bd25-e8a6-4cf7-8efd-1f64df719868 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 689.562198] env[62460]: DEBUG nova.network.neutron [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] [instance: 578627be-e695-4953-8d0e-9763d12b9a28] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.594780] env[62460]: DEBUG nova.compute.manager [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] [instance: 724a8083-2bde-483e-bd5a-a928def284bf] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 689.764259] env[62460]: DEBUG nova.compute.manager [req-9ed9061f-ab73-4a94-bd3b-38c2b4b618c7 req-a2d88909-a4e8-43ae-b7bf-48f309b4fde3 service nova] [instance: 578627be-e695-4953-8d0e-9763d12b9a28] Received event network-vif-deleted-fa6e34fb-aa15-4b5a-909a-4ed21e7e6fbd {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 689.945037] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-241b9bb6-8519-4837-b52e-38d0ae131330 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.952949] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e72d18c0-48bb-426c-9517-f6c31e7be1ad {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.985555] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4e2fa48-39e7-4ef3-95b2-781738449134 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.990766] env[62460]: DEBUG oslo_concurrency.lockutils [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "7e6ff902-4a04-43d5-9014-38c4ec88efc4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 689.990990] env[62460]: DEBUG oslo_concurrency.lockutils [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "7e6ff902-4a04-43d5-9014-38c4ec88efc4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 689.995896] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75ada5af-ea97-4f25-80cf-331d42f8212a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.009292] env[62460]: DEBUG nova.compute.provider_tree [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 690.065111] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] Releasing lock "refresh_cache-578627be-e695-4953-8d0e-9763d12b9a28" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 690.068468] env[62460]: DEBUG nova.compute.manager [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] [instance: 578627be-e695-4953-8d0e-9763d12b9a28] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 690.068468] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] [instance: 578627be-e695-4953-8d0e-9763d12b9a28] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 690.068468] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6af7d4df-c56d-4a5e-87ad-a9f8c26359e9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.081889] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e20bf5e-656b-4e59-834a-9273bf10c35f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.111076] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] [instance: 578627be-e695-4953-8d0e-9763d12b9a28] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 578627be-e695-4953-8d0e-9763d12b9a28 could not be found. [ 690.111312] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] [instance: 578627be-e695-4953-8d0e-9763d12b9a28] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 690.111493] env[62460]: INFO nova.compute.manager [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] [instance: 578627be-e695-4953-8d0e-9763d12b9a28] Took 0.05 seconds to destroy the instance on the hypervisor. [ 690.111743] env[62460]: DEBUG oslo.service.loopingcall [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 690.111955] env[62460]: DEBUG nova.compute.manager [-] [instance: 578627be-e695-4953-8d0e-9763d12b9a28] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 690.112073] env[62460]: DEBUG nova.network.neutron [-] [instance: 578627be-e695-4953-8d0e-9763d12b9a28] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 690.127334] env[62460]: DEBUG nova.network.neutron [-] [instance: 578627be-e695-4953-8d0e-9763d12b9a28] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 690.498599] env[62460]: ERROR nova.compute.manager [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7464bd25-e8a6-4cf7-8efd-1f64df719868, please check neutron logs for more information. [ 690.498599] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 690.498599] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 690.498599] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 690.498599] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 690.498599] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 690.498599] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 690.498599] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 690.498599] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 690.498599] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 690.498599] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 690.498599] env[62460]: ERROR nova.compute.manager raise self.value [ 690.498599] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 690.498599] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 690.498599] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 690.498599] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 690.499108] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 690.499108] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 690.499108] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7464bd25-e8a6-4cf7-8efd-1f64df719868, please check neutron logs for more information. [ 690.499108] env[62460]: ERROR nova.compute.manager [ 690.499108] env[62460]: Traceback (most recent call last): [ 690.499108] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 690.499108] env[62460]: listener.cb(fileno) [ 690.499108] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 690.499108] env[62460]: result = function(*args, **kwargs) [ 690.499108] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 690.499108] env[62460]: return func(*args, **kwargs) [ 690.499108] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 690.499108] env[62460]: raise e [ 690.499108] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 690.499108] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 690.499108] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 690.499108] env[62460]: created_port_ids = self._update_ports_for_instance( [ 690.499108] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 690.499108] env[62460]: with excutils.save_and_reraise_exception(): [ 690.499108] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 690.499108] env[62460]: self.force_reraise() [ 690.499108] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 690.499108] env[62460]: raise self.value [ 690.499108] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 690.499108] env[62460]: updated_port = self._update_port( [ 690.499108] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 690.499108] env[62460]: _ensure_no_port_binding_failure(port) [ 690.499108] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 690.499108] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 690.499978] env[62460]: nova.exception.PortBindingFailed: Binding failed for port 7464bd25-e8a6-4cf7-8efd-1f64df719868, please check neutron logs for more information. [ 690.499978] env[62460]: Removing descriptor: 18 [ 690.512573] env[62460]: DEBUG nova.scheduler.client.report [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 690.606758] env[62460]: DEBUG nova.compute.manager [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] [instance: 724a8083-2bde-483e-bd5a-a928def284bf] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 690.631601] env[62460]: DEBUG nova.virt.hardware [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 690.631827] env[62460]: DEBUG nova.virt.hardware [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 690.632041] env[62460]: DEBUG nova.virt.hardware [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 690.632247] env[62460]: DEBUG nova.virt.hardware [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 690.632403] env[62460]: DEBUG nova.virt.hardware [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 690.632549] env[62460]: DEBUG nova.virt.hardware [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 690.632784] env[62460]: DEBUG nova.virt.hardware [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 690.632918] env[62460]: DEBUG nova.virt.hardware [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 690.633155] env[62460]: DEBUG nova.virt.hardware [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 690.633366] env[62460]: DEBUG nova.virt.hardware [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 690.633607] env[62460]: DEBUG nova.virt.hardware [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 690.633965] env[62460]: DEBUG nova.network.neutron [-] [instance: 578627be-e695-4953-8d0e-9763d12b9a28] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 690.635630] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b797527-6cee-430b-8155-50b8b7aac4bb {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.643776] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b179e6e-6393-469c-9a55-d9e85b044b51 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.658369] env[62460]: ERROR nova.compute.manager [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] [instance: 724a8083-2bde-483e-bd5a-a928def284bf] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7464bd25-e8a6-4cf7-8efd-1f64df719868, please check neutron logs for more information. [ 690.658369] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] Traceback (most recent call last): [ 690.658369] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 690.658369] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] yield resources [ 690.658369] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 690.658369] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] self.driver.spawn(context, instance, image_meta, [ 690.658369] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 690.658369] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 690.658369] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 690.658369] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] vm_ref = self.build_virtual_machine(instance, [ 690.658369] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 690.658687] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] vif_infos = vmwarevif.get_vif_info(self._session, [ 690.658687] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 690.658687] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] for vif in network_info: [ 690.658687] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 690.658687] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] return self._sync_wrapper(fn, *args, **kwargs) [ 690.658687] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 690.658687] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] self.wait() [ 690.658687] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 690.658687] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] self[:] = self._gt.wait() [ 690.658687] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 690.658687] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] return self._exit_event.wait() [ 690.658687] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 690.658687] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] current.throw(*self._exc) [ 690.659254] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 690.659254] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] result = function(*args, **kwargs) [ 690.659254] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 690.659254] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] return func(*args, **kwargs) [ 690.659254] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 690.659254] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] raise e [ 690.659254] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 690.659254] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] nwinfo = self.network_api.allocate_for_instance( [ 690.659254] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 690.659254] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] created_port_ids = self._update_ports_for_instance( [ 690.659254] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 690.659254] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] with excutils.save_and_reraise_exception(): [ 690.659254] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 690.659797] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] self.force_reraise() [ 690.659797] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 690.659797] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] raise self.value [ 690.659797] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 690.659797] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] updated_port = self._update_port( [ 690.659797] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 690.659797] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] _ensure_no_port_binding_failure(port) [ 690.659797] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 690.659797] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] raise exception.PortBindingFailed(port_id=port['id']) [ 690.659797] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] nova.exception.PortBindingFailed: Binding failed for port 7464bd25-e8a6-4cf7-8efd-1f64df719868, please check neutron logs for more information. [ 690.659797] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] [ 690.659797] env[62460]: INFO nova.compute.manager [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] [instance: 724a8083-2bde-483e-bd5a-a928def284bf] Terminating instance [ 690.660654] env[62460]: DEBUG oslo_concurrency.lockutils [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] Acquiring lock "refresh_cache-724a8083-2bde-483e-bd5a-a928def284bf" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 690.660814] env[62460]: DEBUG oslo_concurrency.lockutils [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] Acquired lock "refresh_cache-724a8083-2bde-483e-bd5a-a928def284bf" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 690.660977] env[62460]: DEBUG nova.network.neutron [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] [instance: 724a8083-2bde-483e-bd5a-a928def284bf] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 691.017672] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.435s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 691.018322] env[62460]: DEBUG nova.compute.manager [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 691.020838] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.228s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 691.022276] env[62460]: INFO nova.compute.claims [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 691.139820] env[62460]: INFO nova.compute.manager [-] [instance: 578627be-e695-4953-8d0e-9763d12b9a28] Took 1.03 seconds to deallocate network for instance. [ 691.144886] env[62460]: DEBUG nova.compute.claims [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] [instance: 578627be-e695-4953-8d0e-9763d12b9a28] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 691.145123] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 691.177791] env[62460]: DEBUG nova.network.neutron [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] [instance: 724a8083-2bde-483e-bd5a-a928def284bf] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 691.265909] env[62460]: DEBUG nova.network.neutron [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] [instance: 724a8083-2bde-483e-bd5a-a928def284bf] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 691.526991] env[62460]: DEBUG nova.compute.utils [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 691.532150] env[62460]: DEBUG nova.compute.manager [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 691.532150] env[62460]: DEBUG nova.network.neutron [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 691.582318] env[62460]: DEBUG nova.policy [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b466a2e6d99e42ec9e05cc198a8ada9f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a4d173db47a24afc8fe74b80e364b361', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 691.769202] env[62460]: DEBUG oslo_concurrency.lockutils [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] Releasing lock "refresh_cache-724a8083-2bde-483e-bd5a-a928def284bf" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 691.769679] env[62460]: DEBUG nova.compute.manager [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] [instance: 724a8083-2bde-483e-bd5a-a928def284bf] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 691.769875] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] [instance: 724a8083-2bde-483e-bd5a-a928def284bf] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 691.770208] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-12387f56-a52e-45cb-8e92-8ddbc7e0393a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.785344] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0813b68-c983-47f5-a50f-0ad73d27b924 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.807950] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] [instance: 724a8083-2bde-483e-bd5a-a928def284bf] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 724a8083-2bde-483e-bd5a-a928def284bf could not be found. [ 691.808209] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] [instance: 724a8083-2bde-483e-bd5a-a928def284bf] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 691.808396] env[62460]: INFO nova.compute.manager [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] [instance: 724a8083-2bde-483e-bd5a-a928def284bf] Took 0.04 seconds to destroy the instance on the hypervisor. [ 691.808639] env[62460]: DEBUG oslo.service.loopingcall [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 691.808863] env[62460]: DEBUG nova.compute.manager [-] [instance: 724a8083-2bde-483e-bd5a-a928def284bf] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 691.808969] env[62460]: DEBUG nova.network.neutron [-] [instance: 724a8083-2bde-483e-bd5a-a928def284bf] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 691.821016] env[62460]: DEBUG nova.compute.manager [req-6096bc3f-a326-4deb-bd84-7a72cab9062f req-7a362ff7-72d7-4100-b749-db28e320db59 service nova] [instance: 724a8083-2bde-483e-bd5a-a928def284bf] Received event network-changed-7464bd25-e8a6-4cf7-8efd-1f64df719868 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 691.821016] env[62460]: DEBUG nova.compute.manager [req-6096bc3f-a326-4deb-bd84-7a72cab9062f req-7a362ff7-72d7-4100-b749-db28e320db59 service nova] [instance: 724a8083-2bde-483e-bd5a-a928def284bf] Refreshing instance network info cache due to event network-changed-7464bd25-e8a6-4cf7-8efd-1f64df719868. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 691.821016] env[62460]: DEBUG oslo_concurrency.lockutils [req-6096bc3f-a326-4deb-bd84-7a72cab9062f req-7a362ff7-72d7-4100-b749-db28e320db59 service nova] Acquiring lock "refresh_cache-724a8083-2bde-483e-bd5a-a928def284bf" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 691.821016] env[62460]: DEBUG oslo_concurrency.lockutils [req-6096bc3f-a326-4deb-bd84-7a72cab9062f req-7a362ff7-72d7-4100-b749-db28e320db59 service nova] Acquired lock "refresh_cache-724a8083-2bde-483e-bd5a-a928def284bf" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 691.821016] env[62460]: DEBUG nova.network.neutron [req-6096bc3f-a326-4deb-bd84-7a72cab9062f req-7a362ff7-72d7-4100-b749-db28e320db59 service nova] [instance: 724a8083-2bde-483e-bd5a-a928def284bf] Refreshing network info cache for port 7464bd25-e8a6-4cf7-8efd-1f64df719868 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 691.824027] env[62460]: DEBUG nova.network.neutron [-] [instance: 724a8083-2bde-483e-bd5a-a928def284bf] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 691.914789] env[62460]: DEBUG nova.network.neutron [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] Successfully created port: c4831938-0e88-4552-9ca7-091fbd50dd0f {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 692.031440] env[62460]: DEBUG nova.compute.manager [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 692.325627] env[62460]: DEBUG nova.network.neutron [-] [instance: 724a8083-2bde-483e-bd5a-a928def284bf] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 692.344451] env[62460]: DEBUG nova.network.neutron [req-6096bc3f-a326-4deb-bd84-7a72cab9062f req-7a362ff7-72d7-4100-b749-db28e320db59 service nova] [instance: 724a8083-2bde-483e-bd5a-a928def284bf] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 692.369868] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e289776-47da-483f-93f2-319683ca63e0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.376745] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-584c9e25-e1b2-496d-bbd5-67d617883fa5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.407108] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-157f2764-3a73-4ffc-a0c0-8c82ab1adaa9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.414404] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b1e5d9d-dd42-4870-9147-f81ddae59332 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.428525] env[62460]: DEBUG nova.compute.provider_tree [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 692.606201] env[62460]: DEBUG nova.network.neutron [req-6096bc3f-a326-4deb-bd84-7a72cab9062f req-7a362ff7-72d7-4100-b749-db28e320db59 service nova] [instance: 724a8083-2bde-483e-bd5a-a928def284bf] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 692.827550] env[62460]: INFO nova.compute.manager [-] [instance: 724a8083-2bde-483e-bd5a-a928def284bf] Took 1.02 seconds to deallocate network for instance. [ 692.830072] env[62460]: DEBUG nova.compute.claims [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] [instance: 724a8083-2bde-483e-bd5a-a928def284bf] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 692.830305] env[62460]: DEBUG oslo_concurrency.lockutils [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 692.932433] env[62460]: DEBUG nova.scheduler.client.report [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 693.006624] env[62460]: ERROR nova.compute.manager [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c4831938-0e88-4552-9ca7-091fbd50dd0f, please check neutron logs for more information. [ 693.006624] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 693.006624] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 693.006624] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 693.006624] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 693.006624] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 693.006624] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 693.006624] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 693.006624] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 693.006624] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 693.006624] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 693.006624] env[62460]: ERROR nova.compute.manager raise self.value [ 693.006624] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 693.006624] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 693.006624] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 693.006624] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 693.007203] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 693.007203] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 693.007203] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c4831938-0e88-4552-9ca7-091fbd50dd0f, please check neutron logs for more information. [ 693.007203] env[62460]: ERROR nova.compute.manager [ 693.007203] env[62460]: Traceback (most recent call last): [ 693.007203] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 693.007203] env[62460]: listener.cb(fileno) [ 693.007203] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 693.007203] env[62460]: result = function(*args, **kwargs) [ 693.007203] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 693.007203] env[62460]: return func(*args, **kwargs) [ 693.007203] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 693.007203] env[62460]: raise e [ 693.007203] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 693.007203] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 693.007203] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 693.007203] env[62460]: created_port_ids = self._update_ports_for_instance( [ 693.007203] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 693.007203] env[62460]: with excutils.save_and_reraise_exception(): [ 693.007203] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 693.007203] env[62460]: self.force_reraise() [ 693.007203] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 693.007203] env[62460]: raise self.value [ 693.007203] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 693.007203] env[62460]: updated_port = self._update_port( [ 693.007203] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 693.007203] env[62460]: _ensure_no_port_binding_failure(port) [ 693.007203] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 693.007203] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 693.007977] env[62460]: nova.exception.PortBindingFailed: Binding failed for port c4831938-0e88-4552-9ca7-091fbd50dd0f, please check neutron logs for more information. [ 693.007977] env[62460]: Removing descriptor: 18 [ 693.044514] env[62460]: DEBUG nova.compute.manager [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 693.071801] env[62460]: DEBUG nova.virt.hardware [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 693.072150] env[62460]: DEBUG nova.virt.hardware [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 693.072392] env[62460]: DEBUG nova.virt.hardware [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 693.072615] env[62460]: DEBUG nova.virt.hardware [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 693.072838] env[62460]: DEBUG nova.virt.hardware [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 693.073060] env[62460]: DEBUG nova.virt.hardware [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 693.073349] env[62460]: DEBUG nova.virt.hardware [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 693.073557] env[62460]: DEBUG nova.virt.hardware [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 693.073791] env[62460]: DEBUG nova.virt.hardware [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 693.074038] env[62460]: DEBUG nova.virt.hardware [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 693.074285] env[62460]: DEBUG nova.virt.hardware [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 693.075297] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b78eee32-6956-47e1-b7f1-653ce058627a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.084679] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6a801c8-132f-4288-bcfe-e36c7742cae2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.101577] env[62460]: ERROR nova.compute.manager [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c4831938-0e88-4552-9ca7-091fbd50dd0f, please check neutron logs for more information. [ 693.101577] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] Traceback (most recent call last): [ 693.101577] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 693.101577] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] yield resources [ 693.101577] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 693.101577] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] self.driver.spawn(context, instance, image_meta, [ 693.101577] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 693.101577] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] self._vmops.spawn(context, instance, image_meta, injected_files, [ 693.101577] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 693.101577] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] vm_ref = self.build_virtual_machine(instance, [ 693.101577] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 693.101884] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] vif_infos = vmwarevif.get_vif_info(self._session, [ 693.101884] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 693.101884] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] for vif in network_info: [ 693.101884] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 693.101884] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] return self._sync_wrapper(fn, *args, **kwargs) [ 693.101884] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 693.101884] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] self.wait() [ 693.101884] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 693.101884] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] self[:] = self._gt.wait() [ 693.101884] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 693.101884] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] return self._exit_event.wait() [ 693.101884] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 693.101884] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] current.throw(*self._exc) [ 693.102271] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 693.102271] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] result = function(*args, **kwargs) [ 693.102271] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 693.102271] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] return func(*args, **kwargs) [ 693.102271] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 693.102271] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] raise e [ 693.102271] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 693.102271] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] nwinfo = self.network_api.allocate_for_instance( [ 693.102271] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 693.102271] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] created_port_ids = self._update_ports_for_instance( [ 693.102271] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 693.102271] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] with excutils.save_and_reraise_exception(): [ 693.102271] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 693.102623] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] self.force_reraise() [ 693.102623] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 693.102623] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] raise self.value [ 693.102623] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 693.102623] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] updated_port = self._update_port( [ 693.102623] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 693.102623] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] _ensure_no_port_binding_failure(port) [ 693.102623] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 693.102623] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] raise exception.PortBindingFailed(port_id=port['id']) [ 693.102623] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] nova.exception.PortBindingFailed: Binding failed for port c4831938-0e88-4552-9ca7-091fbd50dd0f, please check neutron logs for more information. [ 693.102623] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] [ 693.102623] env[62460]: INFO nova.compute.manager [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] Terminating instance [ 693.103980] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Acquiring lock "refresh_cache-0d29a4ce-6689-4d0c-8144-e05853eb3a60" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 693.104166] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Acquired lock "refresh_cache-0d29a4ce-6689-4d0c-8144-e05853eb3a60" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 693.104340] env[62460]: DEBUG nova.network.neutron [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 693.108036] env[62460]: DEBUG oslo_concurrency.lockutils [req-6096bc3f-a326-4deb-bd84-7a72cab9062f req-7a362ff7-72d7-4100-b749-db28e320db59 service nova] Releasing lock "refresh_cache-724a8083-2bde-483e-bd5a-a928def284bf" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 693.108264] env[62460]: DEBUG nova.compute.manager [req-6096bc3f-a326-4deb-bd84-7a72cab9062f req-7a362ff7-72d7-4100-b749-db28e320db59 service nova] [instance: 724a8083-2bde-483e-bd5a-a928def284bf] Received event network-vif-deleted-7464bd25-e8a6-4cf7-8efd-1f64df719868 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 693.437387] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.416s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 693.437892] env[62460]: DEBUG nova.compute.manager [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 693.440818] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.755s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 693.442329] env[62460]: INFO nova.compute.claims [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 693.622014] env[62460]: DEBUG nova.network.neutron [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 693.717352] env[62460]: DEBUG nova.network.neutron [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 693.860249] env[62460]: DEBUG nova.compute.manager [req-a547f09c-8f77-48f9-855a-ca2433b32ef4 req-ccf49045-3637-4d60-ad5f-a83021d5e563 service nova] [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] Received event network-changed-c4831938-0e88-4552-9ca7-091fbd50dd0f {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 693.861275] env[62460]: DEBUG nova.compute.manager [req-a547f09c-8f77-48f9-855a-ca2433b32ef4 req-ccf49045-3637-4d60-ad5f-a83021d5e563 service nova] [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] Refreshing instance network info cache due to event network-changed-c4831938-0e88-4552-9ca7-091fbd50dd0f. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 693.861275] env[62460]: DEBUG oslo_concurrency.lockutils [req-a547f09c-8f77-48f9-855a-ca2433b32ef4 req-ccf49045-3637-4d60-ad5f-a83021d5e563 service nova] Acquiring lock "refresh_cache-0d29a4ce-6689-4d0c-8144-e05853eb3a60" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 693.947890] env[62460]: DEBUG nova.compute.utils [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 693.951107] env[62460]: DEBUG nova.compute.manager [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 693.951280] env[62460]: DEBUG nova.network.neutron [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 694.003807] env[62460]: DEBUG nova.policy [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b466a2e6d99e42ec9e05cc198a8ada9f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a4d173db47a24afc8fe74b80e364b361', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 694.221101] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Releasing lock "refresh_cache-0d29a4ce-6689-4d0c-8144-e05853eb3a60" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 694.221101] env[62460]: DEBUG nova.compute.manager [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 694.221101] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 694.222022] env[62460]: DEBUG oslo_concurrency.lockutils [req-a547f09c-8f77-48f9-855a-ca2433b32ef4 req-ccf49045-3637-4d60-ad5f-a83021d5e563 service nova] Acquired lock "refresh_cache-0d29a4ce-6689-4d0c-8144-e05853eb3a60" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 694.222022] env[62460]: DEBUG nova.network.neutron [req-a547f09c-8f77-48f9-855a-ca2433b32ef4 req-ccf49045-3637-4d60-ad5f-a83021d5e563 service nova] [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] Refreshing network info cache for port c4831938-0e88-4552-9ca7-091fbd50dd0f {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 694.222994] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1fdec594-af3b-4b3b-88d5-32bfa46d435b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.233956] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd733195-dd02-411a-850c-145c30f0f8aa {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.255828] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0d29a4ce-6689-4d0c-8144-e05853eb3a60 could not be found. [ 694.255971] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 694.256096] env[62460]: INFO nova.compute.manager [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] Took 0.04 seconds to destroy the instance on the hypervisor. [ 694.256347] env[62460]: DEBUG oslo.service.loopingcall [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 694.256580] env[62460]: DEBUG nova.compute.manager [-] [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 694.256679] env[62460]: DEBUG nova.network.neutron [-] [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 694.263797] env[62460]: DEBUG nova.network.neutron [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] Successfully created port: 8e4514d2-11c6-4bd9-a27f-0b171dad4c9d {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 694.271349] env[62460]: DEBUG nova.network.neutron [-] [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 694.452629] env[62460]: DEBUG nova.compute.manager [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 694.743270] env[62460]: DEBUG nova.network.neutron [req-a547f09c-8f77-48f9-855a-ca2433b32ef4 req-ccf49045-3637-4d60-ad5f-a83021d5e563 service nova] [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 694.773982] env[62460]: DEBUG nova.network.neutron [-] [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 694.796420] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c3aec15-2e86-4ae7-9d5f-42eaf70537f6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.804403] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fd403b1-b650-4f1c-a224-23b9d32b1e6d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.840339] env[62460]: DEBUG nova.network.neutron [req-a547f09c-8f77-48f9-855a-ca2433b32ef4 req-ccf49045-3637-4d60-ad5f-a83021d5e563 service nova] [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 694.841958] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-749ffdad-cf51-47e4-893d-b6e799591ee6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.849986] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4ffeebf-bfdd-4806-a07a-1f894a50f6d6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.864639] env[62460]: DEBUG nova.compute.provider_tree [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 695.252394] env[62460]: ERROR nova.compute.manager [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8e4514d2-11c6-4bd9-a27f-0b171dad4c9d, please check neutron logs for more information. [ 695.252394] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 695.252394] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 695.252394] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 695.252394] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 695.252394] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 695.252394] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 695.252394] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 695.252394] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 695.252394] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 695.252394] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 695.252394] env[62460]: ERROR nova.compute.manager raise self.value [ 695.252394] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 695.252394] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 695.252394] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 695.252394] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 695.253126] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 695.253126] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 695.253126] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8e4514d2-11c6-4bd9-a27f-0b171dad4c9d, please check neutron logs for more information. [ 695.253126] env[62460]: ERROR nova.compute.manager [ 695.253126] env[62460]: Traceback (most recent call last): [ 695.253126] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 695.253126] env[62460]: listener.cb(fileno) [ 695.253126] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 695.253126] env[62460]: result = function(*args, **kwargs) [ 695.253126] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 695.253126] env[62460]: return func(*args, **kwargs) [ 695.253126] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 695.253126] env[62460]: raise e [ 695.253126] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 695.253126] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 695.253126] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 695.253126] env[62460]: created_port_ids = self._update_ports_for_instance( [ 695.253126] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 695.253126] env[62460]: with excutils.save_and_reraise_exception(): [ 695.253126] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 695.253126] env[62460]: self.force_reraise() [ 695.253126] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 695.253126] env[62460]: raise self.value [ 695.253126] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 695.253126] env[62460]: updated_port = self._update_port( [ 695.253126] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 695.253126] env[62460]: _ensure_no_port_binding_failure(port) [ 695.253126] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 695.253126] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 695.253929] env[62460]: nova.exception.PortBindingFailed: Binding failed for port 8e4514d2-11c6-4bd9-a27f-0b171dad4c9d, please check neutron logs for more information. [ 695.253929] env[62460]: Removing descriptor: 18 [ 695.279657] env[62460]: INFO nova.compute.manager [-] [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] Took 1.02 seconds to deallocate network for instance. [ 695.281913] env[62460]: DEBUG nova.compute.claims [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 695.282128] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 695.346111] env[62460]: DEBUG oslo_concurrency.lockutils [req-a547f09c-8f77-48f9-855a-ca2433b32ef4 req-ccf49045-3637-4d60-ad5f-a83021d5e563 service nova] Releasing lock "refresh_cache-0d29a4ce-6689-4d0c-8144-e05853eb3a60" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 695.346387] env[62460]: DEBUG nova.compute.manager [req-a547f09c-8f77-48f9-855a-ca2433b32ef4 req-ccf49045-3637-4d60-ad5f-a83021d5e563 service nova] [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] Received event network-vif-deleted-c4831938-0e88-4552-9ca7-091fbd50dd0f {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 695.367641] env[62460]: DEBUG nova.scheduler.client.report [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 695.465211] env[62460]: DEBUG nova.compute.manager [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 695.489136] env[62460]: DEBUG nova.virt.hardware [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 695.489381] env[62460]: DEBUG nova.virt.hardware [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 695.489539] env[62460]: DEBUG nova.virt.hardware [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 695.489717] env[62460]: DEBUG nova.virt.hardware [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 695.489862] env[62460]: DEBUG nova.virt.hardware [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 695.490016] env[62460]: DEBUG nova.virt.hardware [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 695.490226] env[62460]: DEBUG nova.virt.hardware [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 695.490385] env[62460]: DEBUG nova.virt.hardware [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 695.490547] env[62460]: DEBUG nova.virt.hardware [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 695.490706] env[62460]: DEBUG nova.virt.hardware [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 695.490877] env[62460]: DEBUG nova.virt.hardware [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 695.491728] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9e65102-1678-4c70-8217-f9c5d3b644df {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.499393] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-847228e5-bd83-4ed3-8ad7-466aef11a26e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.512424] env[62460]: ERROR nova.compute.manager [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8e4514d2-11c6-4bd9-a27f-0b171dad4c9d, please check neutron logs for more information. [ 695.512424] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] Traceback (most recent call last): [ 695.512424] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 695.512424] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] yield resources [ 695.512424] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 695.512424] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] self.driver.spawn(context, instance, image_meta, [ 695.512424] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 695.512424] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] self._vmops.spawn(context, instance, image_meta, injected_files, [ 695.512424] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 695.512424] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] vm_ref = self.build_virtual_machine(instance, [ 695.512424] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 695.512776] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] vif_infos = vmwarevif.get_vif_info(self._session, [ 695.512776] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 695.512776] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] for vif in network_info: [ 695.512776] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 695.512776] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] return self._sync_wrapper(fn, *args, **kwargs) [ 695.512776] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 695.512776] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] self.wait() [ 695.512776] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 695.512776] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] self[:] = self._gt.wait() [ 695.512776] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 695.512776] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] return self._exit_event.wait() [ 695.512776] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 695.512776] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] current.throw(*self._exc) [ 695.513118] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 695.513118] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] result = function(*args, **kwargs) [ 695.513118] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 695.513118] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] return func(*args, **kwargs) [ 695.513118] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 695.513118] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] raise e [ 695.513118] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 695.513118] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] nwinfo = self.network_api.allocate_for_instance( [ 695.513118] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 695.513118] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] created_port_ids = self._update_ports_for_instance( [ 695.513118] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 695.513118] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] with excutils.save_and_reraise_exception(): [ 695.513118] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 695.513528] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] self.force_reraise() [ 695.513528] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 695.513528] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] raise self.value [ 695.513528] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 695.513528] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] updated_port = self._update_port( [ 695.513528] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 695.513528] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] _ensure_no_port_binding_failure(port) [ 695.513528] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 695.513528] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] raise exception.PortBindingFailed(port_id=port['id']) [ 695.513528] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] nova.exception.PortBindingFailed: Binding failed for port 8e4514d2-11c6-4bd9-a27f-0b171dad4c9d, please check neutron logs for more information. [ 695.513528] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] [ 695.513528] env[62460]: INFO nova.compute.manager [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] Terminating instance [ 695.514646] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Acquiring lock "refresh_cache-02b7a3c3-bdfd-47d9-acf7-afd7725cea64" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 695.514806] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Acquired lock "refresh_cache-02b7a3c3-bdfd-47d9-acf7-afd7725cea64" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 695.514971] env[62460]: DEBUG nova.network.neutron [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 695.875948] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.435s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 695.876537] env[62460]: DEBUG nova.compute.manager [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 695.879997] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.125s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 695.882060] env[62460]: INFO nova.compute.claims [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 695.928241] env[62460]: DEBUG nova.compute.manager [req-73d3b202-83c9-430a-bf13-bc66bb54c781 req-0e85fd2d-41c2-4c7d-9654-ea54dbae455d service nova] [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] Received event network-changed-8e4514d2-11c6-4bd9-a27f-0b171dad4c9d {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 695.928573] env[62460]: DEBUG nova.compute.manager [req-73d3b202-83c9-430a-bf13-bc66bb54c781 req-0e85fd2d-41c2-4c7d-9654-ea54dbae455d service nova] [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] Refreshing instance network info cache due to event network-changed-8e4514d2-11c6-4bd9-a27f-0b171dad4c9d. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 695.928573] env[62460]: DEBUG oslo_concurrency.lockutils [req-73d3b202-83c9-430a-bf13-bc66bb54c781 req-0e85fd2d-41c2-4c7d-9654-ea54dbae455d service nova] Acquiring lock "refresh_cache-02b7a3c3-bdfd-47d9-acf7-afd7725cea64" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 696.030078] env[62460]: DEBUG nova.network.neutron [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 696.134895] env[62460]: DEBUG nova.network.neutron [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.386224] env[62460]: DEBUG nova.compute.utils [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 696.389509] env[62460]: DEBUG nova.compute.manager [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 696.389696] env[62460]: DEBUG nova.network.neutron [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 696.447672] env[62460]: DEBUG nova.policy [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7372bfba0c624a05bfab97f6a0367d5b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '001591acb4df47a29c7e639b3eba7ad0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 696.638067] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Releasing lock "refresh_cache-02b7a3c3-bdfd-47d9-acf7-afd7725cea64" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 696.638538] env[62460]: DEBUG nova.compute.manager [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 696.638737] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 696.639075] env[62460]: DEBUG oslo_concurrency.lockutils [req-73d3b202-83c9-430a-bf13-bc66bb54c781 req-0e85fd2d-41c2-4c7d-9654-ea54dbae455d service nova] Acquired lock "refresh_cache-02b7a3c3-bdfd-47d9-acf7-afd7725cea64" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.639253] env[62460]: DEBUG nova.network.neutron [req-73d3b202-83c9-430a-bf13-bc66bb54c781 req-0e85fd2d-41c2-4c7d-9654-ea54dbae455d service nova] [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] Refreshing network info cache for port 8e4514d2-11c6-4bd9-a27f-0b171dad4c9d {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 696.640351] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-212e33e9-df50-4e5b-b90d-22e8c1d4a080 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.651397] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6838891e-84c1-4670-b223-ee31cd8f3782 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.672559] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 02b7a3c3-bdfd-47d9-acf7-afd7725cea64 could not be found. [ 696.672833] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 696.673079] env[62460]: INFO nova.compute.manager [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] Took 0.03 seconds to destroy the instance on the hypervisor. [ 696.673377] env[62460]: DEBUG oslo.service.loopingcall [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 696.673643] env[62460]: DEBUG nova.compute.manager [-] [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 696.673784] env[62460]: DEBUG nova.network.neutron [-] [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 696.688560] env[62460]: DEBUG nova.network.neutron [-] [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 696.723283] env[62460]: DEBUG nova.network.neutron [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] Successfully created port: bf96148d-c685-42c5-a715-fec0fd939e37 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 696.890780] env[62460]: DEBUG nova.compute.manager [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 697.163433] env[62460]: DEBUG nova.network.neutron [req-73d3b202-83c9-430a-bf13-bc66bb54c781 req-0e85fd2d-41c2-4c7d-9654-ea54dbae455d service nova] [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 697.193136] env[62460]: DEBUG nova.network.neutron [-] [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.268766] env[62460]: DEBUG nova.network.neutron [req-73d3b202-83c9-430a-bf13-bc66bb54c781 req-0e85fd2d-41c2-4c7d-9654-ea54dbae455d service nova] [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.275909] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7d03de7-d0ad-482e-bcb7-74cd68ac3ccd {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.288391] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d45c08f-e1ab-4faa-bda8-5fda6e4ea326 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.319530] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b13008e4-a1c5-4e89-8088-a2daecad2498 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.326608] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ad58184-4ba9-41a5-b975-8232ccc229bb {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.341257] env[62460]: DEBUG nova.compute.provider_tree [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 697.698152] env[62460]: ERROR nova.compute.manager [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port bf96148d-c685-42c5-a715-fec0fd939e37, please check neutron logs for more information. [ 697.698152] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 697.698152] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 697.698152] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 697.698152] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 697.698152] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 697.698152] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 697.698152] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 697.698152] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 697.698152] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 697.698152] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 697.698152] env[62460]: ERROR nova.compute.manager raise self.value [ 697.698152] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 697.698152] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 697.698152] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 697.698152] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 697.698787] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 697.698787] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 697.698787] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port bf96148d-c685-42c5-a715-fec0fd939e37, please check neutron logs for more information. [ 697.698787] env[62460]: ERROR nova.compute.manager [ 697.698787] env[62460]: Traceback (most recent call last): [ 697.698787] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 697.698787] env[62460]: listener.cb(fileno) [ 697.698787] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 697.698787] env[62460]: result = function(*args, **kwargs) [ 697.698787] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 697.698787] env[62460]: return func(*args, **kwargs) [ 697.698787] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 697.698787] env[62460]: raise e [ 697.698787] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 697.698787] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 697.698787] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 697.698787] env[62460]: created_port_ids = self._update_ports_for_instance( [ 697.698787] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 697.698787] env[62460]: with excutils.save_and_reraise_exception(): [ 697.698787] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 697.698787] env[62460]: self.force_reraise() [ 697.698787] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 697.698787] env[62460]: raise self.value [ 697.698787] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 697.698787] env[62460]: updated_port = self._update_port( [ 697.698787] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 697.698787] env[62460]: _ensure_no_port_binding_failure(port) [ 697.698787] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 697.698787] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 697.700142] env[62460]: nova.exception.PortBindingFailed: Binding failed for port bf96148d-c685-42c5-a715-fec0fd939e37, please check neutron logs for more information. [ 697.700142] env[62460]: Removing descriptor: 18 [ 697.700142] env[62460]: INFO nova.compute.manager [-] [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] Took 1.02 seconds to deallocate network for instance. [ 697.701512] env[62460]: DEBUG nova.compute.claims [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 697.701695] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 697.771602] env[62460]: DEBUG oslo_concurrency.lockutils [req-73d3b202-83c9-430a-bf13-bc66bb54c781 req-0e85fd2d-41c2-4c7d-9654-ea54dbae455d service nova] Releasing lock "refresh_cache-02b7a3c3-bdfd-47d9-acf7-afd7725cea64" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 697.772234] env[62460]: DEBUG nova.compute.manager [req-73d3b202-83c9-430a-bf13-bc66bb54c781 req-0e85fd2d-41c2-4c7d-9654-ea54dbae455d service nova] [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] Received event network-vif-deleted-8e4514d2-11c6-4bd9-a27f-0b171dad4c9d {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 697.844296] env[62460]: DEBUG nova.scheduler.client.report [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 697.903173] env[62460]: DEBUG nova.compute.manager [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 697.926663] env[62460]: DEBUG nova.virt.hardware [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 697.926903] env[62460]: DEBUG nova.virt.hardware [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 697.927079] env[62460]: DEBUG nova.virt.hardware [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 697.927271] env[62460]: DEBUG nova.virt.hardware [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 697.927419] env[62460]: DEBUG nova.virt.hardware [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 697.927563] env[62460]: DEBUG nova.virt.hardware [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 697.927766] env[62460]: DEBUG nova.virt.hardware [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 697.927925] env[62460]: DEBUG nova.virt.hardware [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 697.928104] env[62460]: DEBUG nova.virt.hardware [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 697.928272] env[62460]: DEBUG nova.virt.hardware [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 697.928445] env[62460]: DEBUG nova.virt.hardware [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 697.929321] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf1830ee-5ba8-4323-a1f5-12950621edc4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.937271] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4177496-e7d0-47dc-b225-68e0f0c03399 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.952621] env[62460]: ERROR nova.compute.manager [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port bf96148d-c685-42c5-a715-fec0fd939e37, please check neutron logs for more information. [ 697.952621] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] Traceback (most recent call last): [ 697.952621] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 697.952621] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] yield resources [ 697.952621] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 697.952621] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] self.driver.spawn(context, instance, image_meta, [ 697.952621] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 697.952621] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 697.952621] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 697.952621] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] vm_ref = self.build_virtual_machine(instance, [ 697.952621] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 697.953015] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] vif_infos = vmwarevif.get_vif_info(self._session, [ 697.953015] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 697.953015] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] for vif in network_info: [ 697.953015] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 697.953015] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] return self._sync_wrapper(fn, *args, **kwargs) [ 697.953015] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 697.953015] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] self.wait() [ 697.953015] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 697.953015] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] self[:] = self._gt.wait() [ 697.953015] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 697.953015] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] return self._exit_event.wait() [ 697.953015] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 697.953015] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] current.throw(*self._exc) [ 697.953432] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 697.953432] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] result = function(*args, **kwargs) [ 697.953432] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 697.953432] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] return func(*args, **kwargs) [ 697.953432] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 697.953432] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] raise e [ 697.953432] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 697.953432] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] nwinfo = self.network_api.allocate_for_instance( [ 697.953432] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 697.953432] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] created_port_ids = self._update_ports_for_instance( [ 697.953432] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 697.953432] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] with excutils.save_and_reraise_exception(): [ 697.953432] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 697.953816] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] self.force_reraise() [ 697.953816] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 697.953816] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] raise self.value [ 697.953816] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 697.953816] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] updated_port = self._update_port( [ 697.953816] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 697.953816] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] _ensure_no_port_binding_failure(port) [ 697.953816] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 697.953816] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] raise exception.PortBindingFailed(port_id=port['id']) [ 697.953816] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] nova.exception.PortBindingFailed: Binding failed for port bf96148d-c685-42c5-a715-fec0fd939e37, please check neutron logs for more information. [ 697.953816] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] [ 697.953816] env[62460]: INFO nova.compute.manager [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] Terminating instance [ 697.955395] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] Acquiring lock "refresh_cache-47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 697.955623] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] Acquired lock "refresh_cache-47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 697.955807] env[62460]: DEBUG nova.network.neutron [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 697.991746] env[62460]: DEBUG nova.compute.manager [req-dc372339-d316-45f0-afe6-f62d82c8480d req-15ce46f0-8bf3-4794-820f-21626ed7cd1e service nova] [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] Received event network-changed-bf96148d-c685-42c5-a715-fec0fd939e37 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 697.991944] env[62460]: DEBUG nova.compute.manager [req-dc372339-d316-45f0-afe6-f62d82c8480d req-15ce46f0-8bf3-4794-820f-21626ed7cd1e service nova] [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] Refreshing instance network info cache due to event network-changed-bf96148d-c685-42c5-a715-fec0fd939e37. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 697.992197] env[62460]: DEBUG oslo_concurrency.lockutils [req-dc372339-d316-45f0-afe6-f62d82c8480d req-15ce46f0-8bf3-4794-820f-21626ed7cd1e service nova] Acquiring lock "refresh_cache-47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 698.348922] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.469s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 698.349482] env[62460]: DEBUG nova.compute.manager [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 698.352403] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.538s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 698.353813] env[62460]: INFO nova.compute.claims [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 698.475070] env[62460]: DEBUG nova.network.neutron [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 698.568644] env[62460]: DEBUG nova.network.neutron [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.858477] env[62460]: DEBUG nova.compute.utils [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 698.861665] env[62460]: DEBUG nova.compute.manager [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 698.861843] env[62460]: DEBUG nova.network.neutron [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 698.899106] env[62460]: DEBUG nova.policy [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '810ffd6bd798494cb98cdff99387f3f0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '17fe89d746cd45458ab11731a253c25d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 699.071013] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] Releasing lock "refresh_cache-47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 699.071497] env[62460]: DEBUG nova.compute.manager [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 699.071828] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 699.072130] env[62460]: DEBUG oslo_concurrency.lockutils [req-dc372339-d316-45f0-afe6-f62d82c8480d req-15ce46f0-8bf3-4794-820f-21626ed7cd1e service nova] Acquired lock "refresh_cache-47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 699.072290] env[62460]: DEBUG nova.network.neutron [req-dc372339-d316-45f0-afe6-f62d82c8480d req-15ce46f0-8bf3-4794-820f-21626ed7cd1e service nova] [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] Refreshing network info cache for port bf96148d-c685-42c5-a715-fec0fd939e37 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 699.073404] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5fdc4a94-835a-4d38-a03b-4c5a32b92a33 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.082709] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5c9d6b0-b7be-443c-a282-e10b787d9a65 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.107010] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4 could not be found. [ 699.108660] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 699.108660] env[62460]: INFO nova.compute.manager [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] Took 0.04 seconds to destroy the instance on the hypervisor. [ 699.108660] env[62460]: DEBUG oslo.service.loopingcall [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 699.108660] env[62460]: DEBUG nova.compute.manager [-] [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 699.108660] env[62460]: DEBUG nova.network.neutron [-] [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 699.129263] env[62460]: DEBUG nova.network.neutron [-] [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 699.213848] env[62460]: DEBUG nova.network.neutron [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] Successfully created port: b3b66865-1822-47a2-a3d1-45a61b73a2cf {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 699.368068] env[62460]: DEBUG nova.compute.manager [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 699.591220] env[62460]: DEBUG nova.network.neutron [req-dc372339-d316-45f0-afe6-f62d82c8480d req-15ce46f0-8bf3-4794-820f-21626ed7cd1e service nova] [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 699.631451] env[62460]: DEBUG nova.network.neutron [-] [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 699.709901] env[62460]: DEBUG nova.network.neutron [req-dc372339-d316-45f0-afe6-f62d82c8480d req-15ce46f0-8bf3-4794-820f-21626ed7cd1e service nova] [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 699.729196] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0cbf464-5144-4799-8bab-8470ab51f35b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.741536] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0f92649-4f70-4b5d-a595-96f30e742942 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.767772] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8aba62f-c071-46ae-853b-ce59953531da {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.775390] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f4eab0f-cc85-482c-8cee-bfe19fd3f5aa {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.788599] env[62460]: DEBUG nova.compute.provider_tree [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 700.073943] env[62460]: DEBUG nova.compute.manager [req-5d902a96-fb76-44f0-ba1e-be516a0feab0 req-bd8b2151-ac5c-481b-9de5-41d223e3216a service nova] [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] Received event network-changed-b3b66865-1822-47a2-a3d1-45a61b73a2cf {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 700.073943] env[62460]: DEBUG nova.compute.manager [req-5d902a96-fb76-44f0-ba1e-be516a0feab0 req-bd8b2151-ac5c-481b-9de5-41d223e3216a service nova] [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] Refreshing instance network info cache due to event network-changed-b3b66865-1822-47a2-a3d1-45a61b73a2cf. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 700.073943] env[62460]: DEBUG oslo_concurrency.lockutils [req-5d902a96-fb76-44f0-ba1e-be516a0feab0 req-bd8b2151-ac5c-481b-9de5-41d223e3216a service nova] Acquiring lock "refresh_cache-770b6195-ab31-4df0-84b0-e8382732ec32" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 700.073943] env[62460]: DEBUG oslo_concurrency.lockutils [req-5d902a96-fb76-44f0-ba1e-be516a0feab0 req-bd8b2151-ac5c-481b-9de5-41d223e3216a service nova] Acquired lock "refresh_cache-770b6195-ab31-4df0-84b0-e8382732ec32" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 700.074438] env[62460]: DEBUG nova.network.neutron [req-5d902a96-fb76-44f0-ba1e-be516a0feab0 req-bd8b2151-ac5c-481b-9de5-41d223e3216a service nova] [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] Refreshing network info cache for port b3b66865-1822-47a2-a3d1-45a61b73a2cf {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 700.133885] env[62460]: INFO nova.compute.manager [-] [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] Took 1.03 seconds to deallocate network for instance. [ 700.136593] env[62460]: DEBUG nova.compute.claims [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 700.136775] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 700.183007] env[62460]: ERROR nova.compute.manager [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b3b66865-1822-47a2-a3d1-45a61b73a2cf, please check neutron logs for more information. [ 700.183007] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 700.183007] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 700.183007] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 700.183007] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 700.183007] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 700.183007] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 700.183007] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 700.183007] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 700.183007] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 700.183007] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 700.183007] env[62460]: ERROR nova.compute.manager raise self.value [ 700.183007] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 700.183007] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 700.183007] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 700.183007] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 700.183486] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 700.183486] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 700.183486] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b3b66865-1822-47a2-a3d1-45a61b73a2cf, please check neutron logs for more information. [ 700.183486] env[62460]: ERROR nova.compute.manager [ 700.183486] env[62460]: Traceback (most recent call last): [ 700.183486] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 700.183486] env[62460]: listener.cb(fileno) [ 700.183486] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 700.183486] env[62460]: result = function(*args, **kwargs) [ 700.183486] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 700.183486] env[62460]: return func(*args, **kwargs) [ 700.183486] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 700.183486] env[62460]: raise e [ 700.183486] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 700.183486] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 700.183486] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 700.183486] env[62460]: created_port_ids = self._update_ports_for_instance( [ 700.183486] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 700.183486] env[62460]: with excutils.save_and_reraise_exception(): [ 700.183486] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 700.183486] env[62460]: self.force_reraise() [ 700.183486] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 700.183486] env[62460]: raise self.value [ 700.183486] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 700.183486] env[62460]: updated_port = self._update_port( [ 700.183486] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 700.183486] env[62460]: _ensure_no_port_binding_failure(port) [ 700.183486] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 700.183486] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 700.184344] env[62460]: nova.exception.PortBindingFailed: Binding failed for port b3b66865-1822-47a2-a3d1-45a61b73a2cf, please check neutron logs for more information. [ 700.184344] env[62460]: Removing descriptor: 18 [ 700.212778] env[62460]: DEBUG oslo_concurrency.lockutils [req-dc372339-d316-45f0-afe6-f62d82c8480d req-15ce46f0-8bf3-4794-820f-21626ed7cd1e service nova] Releasing lock "refresh_cache-47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 700.213177] env[62460]: DEBUG nova.compute.manager [req-dc372339-d316-45f0-afe6-f62d82c8480d req-15ce46f0-8bf3-4794-820f-21626ed7cd1e service nova] [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] Received event network-vif-deleted-bf96148d-c685-42c5-a715-fec0fd939e37 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 700.291802] env[62460]: DEBUG nova.scheduler.client.report [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 700.382030] env[62460]: DEBUG nova.compute.manager [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 700.409114] env[62460]: DEBUG nova.virt.hardware [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 700.409426] env[62460]: DEBUG nova.virt.hardware [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 700.409642] env[62460]: DEBUG nova.virt.hardware [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 700.409878] env[62460]: DEBUG nova.virt.hardware [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 700.410094] env[62460]: DEBUG nova.virt.hardware [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 700.410276] env[62460]: DEBUG nova.virt.hardware [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 700.410444] env[62460]: DEBUG nova.virt.hardware [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 700.410602] env[62460]: DEBUG nova.virt.hardware [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 700.410767] env[62460]: DEBUG nova.virt.hardware [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 700.410926] env[62460]: DEBUG nova.virt.hardware [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 700.411116] env[62460]: DEBUG nova.virt.hardware [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 700.412082] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc5cbcc1-c298-4d53-804b-f100fcbf4b10 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.419947] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-800ad3d0-f642-438f-be09-de8f3428c1fe {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.433574] env[62460]: ERROR nova.compute.manager [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b3b66865-1822-47a2-a3d1-45a61b73a2cf, please check neutron logs for more information. [ 700.433574] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] Traceback (most recent call last): [ 700.433574] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 700.433574] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] yield resources [ 700.433574] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 700.433574] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] self.driver.spawn(context, instance, image_meta, [ 700.433574] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 700.433574] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] self._vmops.spawn(context, instance, image_meta, injected_files, [ 700.433574] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 700.433574] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] vm_ref = self.build_virtual_machine(instance, [ 700.433574] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 700.433939] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] vif_infos = vmwarevif.get_vif_info(self._session, [ 700.433939] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 700.433939] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] for vif in network_info: [ 700.433939] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 700.433939] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] return self._sync_wrapper(fn, *args, **kwargs) [ 700.433939] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 700.433939] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] self.wait() [ 700.433939] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 700.433939] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] self[:] = self._gt.wait() [ 700.433939] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 700.433939] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] return self._exit_event.wait() [ 700.433939] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 700.433939] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] current.throw(*self._exc) [ 700.434360] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 700.434360] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] result = function(*args, **kwargs) [ 700.434360] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 700.434360] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] return func(*args, **kwargs) [ 700.434360] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 700.434360] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] raise e [ 700.434360] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 700.434360] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] nwinfo = self.network_api.allocate_for_instance( [ 700.434360] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 700.434360] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] created_port_ids = self._update_ports_for_instance( [ 700.434360] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 700.434360] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] with excutils.save_and_reraise_exception(): [ 700.434360] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 700.434740] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] self.force_reraise() [ 700.434740] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 700.434740] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] raise self.value [ 700.434740] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 700.434740] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] updated_port = self._update_port( [ 700.434740] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 700.434740] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] _ensure_no_port_binding_failure(port) [ 700.434740] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 700.434740] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] raise exception.PortBindingFailed(port_id=port['id']) [ 700.434740] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] nova.exception.PortBindingFailed: Binding failed for port b3b66865-1822-47a2-a3d1-45a61b73a2cf, please check neutron logs for more information. [ 700.434740] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] [ 700.434740] env[62460]: INFO nova.compute.manager [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] Terminating instance [ 700.437649] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] Acquiring lock "refresh_cache-770b6195-ab31-4df0-84b0-e8382732ec32" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 700.590555] env[62460]: DEBUG nova.network.neutron [req-5d902a96-fb76-44f0-ba1e-be516a0feab0 req-bd8b2151-ac5c-481b-9de5-41d223e3216a service nova] [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 700.656891] env[62460]: DEBUG nova.network.neutron [req-5d902a96-fb76-44f0-ba1e-be516a0feab0 req-bd8b2151-ac5c-481b-9de5-41d223e3216a service nova] [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 700.799219] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.447s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 700.799775] env[62460]: DEBUG nova.compute.manager [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 700.802243] env[62460]: DEBUG oslo_concurrency.lockutils [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.171s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 701.159356] env[62460]: DEBUG oslo_concurrency.lockutils [req-5d902a96-fb76-44f0-ba1e-be516a0feab0 req-bd8b2151-ac5c-481b-9de5-41d223e3216a service nova] Releasing lock "refresh_cache-770b6195-ab31-4df0-84b0-e8382732ec32" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 701.159888] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] Acquired lock "refresh_cache-770b6195-ab31-4df0-84b0-e8382732ec32" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.160165] env[62460]: DEBUG nova.network.neutron [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 701.308701] env[62460]: DEBUG nova.compute.utils [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 701.310200] env[62460]: DEBUG nova.compute.manager [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 701.310383] env[62460]: DEBUG nova.network.neutron [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 701.359248] env[62460]: DEBUG nova.policy [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0ed401cfbf6c48bfa68d6313274accd5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '08f6f6ae3d514d6ca62b665be9034f8e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 701.618032] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddd3cb3e-bd96-44e3-9c39-70fb47ca5fe0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.627509] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-880d679e-860d-40c9-b9fc-0b376a7884cf {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.660312] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76d34761-b356-4640-a59c-37dc45d2c8ee {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.669647] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70aff737-2cb8-4c46-b643-c74c73efa19b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.674186] env[62460]: DEBUG nova.network.neutron [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] Successfully created port: 0c0c5547-5634-45e8-b02c-5c73c6b96852 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 701.685486] env[62460]: DEBUG nova.compute.provider_tree [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 701.687318] env[62460]: DEBUG nova.network.neutron [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 701.778844] env[62460]: DEBUG nova.network.neutron [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.813013] env[62460]: DEBUG nova.compute.manager [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 702.100555] env[62460]: DEBUG nova.compute.manager [req-dad4eaab-1444-4705-b365-7ca983c70566 req-52bf3170-53e6-4fb3-982b-a7d6fc1a42f5 service nova] [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] Received event network-vif-deleted-b3b66865-1822-47a2-a3d1-45a61b73a2cf {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 702.190760] env[62460]: DEBUG nova.scheduler.client.report [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 702.285219] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] Releasing lock "refresh_cache-770b6195-ab31-4df0-84b0-e8382732ec32" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 702.285576] env[62460]: DEBUG nova.compute.manager [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 702.285778] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 702.286099] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d4262287-3c91-46b3-935d-ebc0e47b9e5f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.300365] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf2aee0d-009a-4e3f-837b-9e8683965ce6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.323614] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 770b6195-ab31-4df0-84b0-e8382732ec32 could not be found. [ 702.326701] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 702.326701] env[62460]: INFO nova.compute.manager [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] Took 0.04 seconds to destroy the instance on the hypervisor. [ 702.326701] env[62460]: DEBUG oslo.service.loopingcall [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 702.326701] env[62460]: DEBUG nova.compute.manager [-] [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 702.326701] env[62460]: DEBUG nova.network.neutron [-] [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 702.341272] env[62460]: DEBUG nova.network.neutron [-] [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 702.695214] env[62460]: DEBUG oslo_concurrency.lockutils [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.893s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 702.695908] env[62460]: ERROR nova.compute.manager [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1f0b1f33-f610-4a68-a5b9-d339068ba1bc, please check neutron logs for more information. [ 702.695908] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] Traceback (most recent call last): [ 702.695908] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 702.695908] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] self.driver.spawn(context, instance, image_meta, [ 702.695908] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 702.695908] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 702.695908] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 702.695908] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] vm_ref = self.build_virtual_machine(instance, [ 702.695908] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 702.695908] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] vif_infos = vmwarevif.get_vif_info(self._session, [ 702.695908] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 702.696260] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] for vif in network_info: [ 702.696260] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 702.696260] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] return self._sync_wrapper(fn, *args, **kwargs) [ 702.696260] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 702.696260] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] self.wait() [ 702.696260] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 702.696260] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] self[:] = self._gt.wait() [ 702.696260] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 702.696260] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] return self._exit_event.wait() [ 702.696260] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 702.696260] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] result = hub.switch() [ 702.696260] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 702.696260] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] return self.greenlet.switch() [ 702.696649] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 702.696649] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] result = function(*args, **kwargs) [ 702.696649] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 702.696649] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] return func(*args, **kwargs) [ 702.696649] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 702.696649] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] raise e [ 702.696649] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 702.696649] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] nwinfo = self.network_api.allocate_for_instance( [ 702.696649] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 702.696649] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] created_port_ids = self._update_ports_for_instance( [ 702.696649] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 702.696649] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] with excutils.save_and_reraise_exception(): [ 702.696649] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 702.697028] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] self.force_reraise() [ 702.697028] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 702.697028] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] raise self.value [ 702.697028] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 702.697028] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] updated_port = self._update_port( [ 702.697028] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 702.697028] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] _ensure_no_port_binding_failure(port) [ 702.697028] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 702.697028] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] raise exception.PortBindingFailed(port_id=port['id']) [ 702.697028] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] nova.exception.PortBindingFailed: Binding failed for port 1f0b1f33-f610-4a68-a5b9-d339068ba1bc, please check neutron logs for more information. [ 702.697028] env[62460]: ERROR nova.compute.manager [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] [ 702.697309] env[62460]: DEBUG nova.compute.utils [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] Binding failed for port 1f0b1f33-f610-4a68-a5b9-d339068ba1bc, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 702.697997] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.618s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 702.701431] env[62460]: DEBUG nova.compute.manager [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] Build of instance 252df6ad-e29d-4596-bc99-4aae144bbcc5 was re-scheduled: Binding failed for port 1f0b1f33-f610-4a68-a5b9-d339068ba1bc, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 702.701910] env[62460]: DEBUG nova.compute.manager [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 702.702177] env[62460]: DEBUG oslo_concurrency.lockutils [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] Acquiring lock "refresh_cache-252df6ad-e29d-4596-bc99-4aae144bbcc5" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 702.702363] env[62460]: DEBUG oslo_concurrency.lockutils [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] Acquired lock "refresh_cache-252df6ad-e29d-4596-bc99-4aae144bbcc5" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 702.702553] env[62460]: DEBUG nova.network.neutron [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 702.763858] env[62460]: ERROR nova.compute.manager [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0c0c5547-5634-45e8-b02c-5c73c6b96852, please check neutron logs for more information. [ 702.763858] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 702.763858] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 702.763858] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 702.763858] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 702.763858] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 702.763858] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 702.763858] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 702.763858] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 702.763858] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 702.763858] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 702.763858] env[62460]: ERROR nova.compute.manager raise self.value [ 702.763858] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 702.763858] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 702.763858] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 702.763858] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 702.764382] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 702.764382] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 702.764382] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0c0c5547-5634-45e8-b02c-5c73c6b96852, please check neutron logs for more information. [ 702.764382] env[62460]: ERROR nova.compute.manager [ 702.764382] env[62460]: Traceback (most recent call last): [ 702.764382] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 702.764382] env[62460]: listener.cb(fileno) [ 702.764382] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 702.764382] env[62460]: result = function(*args, **kwargs) [ 702.764382] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 702.764382] env[62460]: return func(*args, **kwargs) [ 702.764382] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 702.764382] env[62460]: raise e [ 702.764382] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 702.764382] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 702.764382] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 702.764382] env[62460]: created_port_ids = self._update_ports_for_instance( [ 702.764382] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 702.764382] env[62460]: with excutils.save_and_reraise_exception(): [ 702.764382] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 702.764382] env[62460]: self.force_reraise() [ 702.764382] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 702.764382] env[62460]: raise self.value [ 702.764382] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 702.764382] env[62460]: updated_port = self._update_port( [ 702.764382] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 702.764382] env[62460]: _ensure_no_port_binding_failure(port) [ 702.764382] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 702.764382] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 702.765304] env[62460]: nova.exception.PortBindingFailed: Binding failed for port 0c0c5547-5634-45e8-b02c-5c73c6b96852, please check neutron logs for more information. [ 702.765304] env[62460]: Removing descriptor: 18 [ 702.826591] env[62460]: DEBUG nova.compute.manager [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 702.844217] env[62460]: DEBUG nova.network.neutron [-] [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.860733] env[62460]: DEBUG nova.virt.hardware [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 702.860979] env[62460]: DEBUG nova.virt.hardware [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 702.861156] env[62460]: DEBUG nova.virt.hardware [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 702.861342] env[62460]: DEBUG nova.virt.hardware [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 702.861490] env[62460]: DEBUG nova.virt.hardware [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 702.861638] env[62460]: DEBUG nova.virt.hardware [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 702.861842] env[62460]: DEBUG nova.virt.hardware [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 702.861996] env[62460]: DEBUG nova.virt.hardware [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 702.862191] env[62460]: DEBUG nova.virt.hardware [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 702.862409] env[62460]: DEBUG nova.virt.hardware [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 702.862588] env[62460]: DEBUG nova.virt.hardware [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 702.863667] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56857db5-43b0-40ed-bb3f-339c3aacf23d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.871511] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d710da8-c4f3-4d5e-9056-1d22f7a6cae0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.885997] env[62460]: ERROR nova.compute.manager [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0c0c5547-5634-45e8-b02c-5c73c6b96852, please check neutron logs for more information. [ 702.885997] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] Traceback (most recent call last): [ 702.885997] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 702.885997] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] yield resources [ 702.885997] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 702.885997] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] self.driver.spawn(context, instance, image_meta, [ 702.885997] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 702.885997] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] self._vmops.spawn(context, instance, image_meta, injected_files, [ 702.885997] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 702.885997] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] vm_ref = self.build_virtual_machine(instance, [ 702.885997] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 702.886409] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] vif_infos = vmwarevif.get_vif_info(self._session, [ 702.886409] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 702.886409] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] for vif in network_info: [ 702.886409] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 702.886409] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] return self._sync_wrapper(fn, *args, **kwargs) [ 702.886409] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 702.886409] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] self.wait() [ 702.886409] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 702.886409] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] self[:] = self._gt.wait() [ 702.886409] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 702.886409] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] return self._exit_event.wait() [ 702.886409] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 702.886409] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] current.throw(*self._exc) [ 702.886803] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 702.886803] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] result = function(*args, **kwargs) [ 702.886803] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 702.886803] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] return func(*args, **kwargs) [ 702.886803] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 702.886803] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] raise e [ 702.886803] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 702.886803] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] nwinfo = self.network_api.allocate_for_instance( [ 702.886803] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 702.886803] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] created_port_ids = self._update_ports_for_instance( [ 702.886803] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 702.886803] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] with excutils.save_and_reraise_exception(): [ 702.886803] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 702.887211] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] self.force_reraise() [ 702.887211] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 702.887211] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] raise self.value [ 702.887211] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 702.887211] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] updated_port = self._update_port( [ 702.887211] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 702.887211] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] _ensure_no_port_binding_failure(port) [ 702.887211] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 702.887211] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] raise exception.PortBindingFailed(port_id=port['id']) [ 702.887211] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] nova.exception.PortBindingFailed: Binding failed for port 0c0c5547-5634-45e8-b02c-5c73c6b96852, please check neutron logs for more information. [ 702.887211] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] [ 702.887211] env[62460]: INFO nova.compute.manager [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] Terminating instance [ 702.887986] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] Acquiring lock "refresh_cache-2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 702.888157] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] Acquired lock "refresh_cache-2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 702.888322] env[62460]: DEBUG nova.network.neutron [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 703.222394] env[62460]: DEBUG nova.network.neutron [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 703.289681] env[62460]: DEBUG nova.network.neutron [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.347255] env[62460]: INFO nova.compute.manager [-] [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] Took 1.02 seconds to deallocate network for instance. [ 703.349370] env[62460]: DEBUG nova.compute.claims [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 703.349549] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 703.417213] env[62460]: DEBUG nova.network.neutron [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 703.487337] env[62460]: DEBUG nova.network.neutron [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.501456] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8697a3c3-7375-4761-8790-8ddc11f47047 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.509434] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0be7615-a66a-4ee5-9d0e-94e7a3193b0e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.538402] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c1efa1e-bc3d-4d70-9d64-356e96c50ba7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.545673] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-696b4bea-9b88-434d-9050-9690a6e21230 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.558543] env[62460]: DEBUG nova.compute.provider_tree [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 703.792485] env[62460]: DEBUG oslo_concurrency.lockutils [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] Releasing lock "refresh_cache-252df6ad-e29d-4596-bc99-4aae144bbcc5" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 703.792741] env[62460]: DEBUG nova.compute.manager [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 703.792912] env[62460]: DEBUG nova.compute.manager [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 703.793098] env[62460]: DEBUG nova.network.neutron [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 703.806918] env[62460]: DEBUG nova.network.neutron [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 703.990039] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] Releasing lock "refresh_cache-2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 703.990493] env[62460]: DEBUG nova.compute.manager [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 703.990736] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 703.991082] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b6834587-d7e0-4115-a47e-067de5054e96 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.000826] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-affedc18-e51d-48db-8227-e0b658b3c86b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.021643] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae could not be found. [ 704.021845] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 704.022038] env[62460]: INFO nova.compute.manager [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] Took 0.03 seconds to destroy the instance on the hypervisor. [ 704.022360] env[62460]: DEBUG oslo.service.loopingcall [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 704.022576] env[62460]: DEBUG nova.compute.manager [-] [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 704.022672] env[62460]: DEBUG nova.network.neutron [-] [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 704.036843] env[62460]: DEBUG nova.network.neutron [-] [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 704.061784] env[62460]: DEBUG nova.scheduler.client.report [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 704.125278] env[62460]: DEBUG nova.compute.manager [req-b9c40289-7f61-4cc3-828f-f54eb0cc2396 req-f6771c48-2ded-4d40-915f-47c574769f60 service nova] [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] Received event network-changed-0c0c5547-5634-45e8-b02c-5c73c6b96852 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 704.125553] env[62460]: DEBUG nova.compute.manager [req-b9c40289-7f61-4cc3-828f-f54eb0cc2396 req-f6771c48-2ded-4d40-915f-47c574769f60 service nova] [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] Refreshing instance network info cache due to event network-changed-0c0c5547-5634-45e8-b02c-5c73c6b96852. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 704.125686] env[62460]: DEBUG oslo_concurrency.lockutils [req-b9c40289-7f61-4cc3-828f-f54eb0cc2396 req-f6771c48-2ded-4d40-915f-47c574769f60 service nova] Acquiring lock "refresh_cache-2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 704.125826] env[62460]: DEBUG oslo_concurrency.lockutils [req-b9c40289-7f61-4cc3-828f-f54eb0cc2396 req-f6771c48-2ded-4d40-915f-47c574769f60 service nova] Acquired lock "refresh_cache-2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 704.125978] env[62460]: DEBUG nova.network.neutron [req-b9c40289-7f61-4cc3-828f-f54eb0cc2396 req-f6771c48-2ded-4d40-915f-47c574769f60 service nova] [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] Refreshing network info cache for port 0c0c5547-5634-45e8-b02c-5c73c6b96852 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 704.310206] env[62460]: DEBUG nova.network.neutron [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.539496] env[62460]: DEBUG nova.network.neutron [-] [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.566615] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.869s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 704.567254] env[62460]: ERROR nova.compute.manager [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 956ced4a-8d22-4660-b90c-a7a197dd8adb, please check neutron logs for more information. [ 704.567254] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] Traceback (most recent call last): [ 704.567254] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 704.567254] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] self.driver.spawn(context, instance, image_meta, [ 704.567254] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 704.567254] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 704.567254] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 704.567254] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] vm_ref = self.build_virtual_machine(instance, [ 704.567254] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 704.567254] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] vif_infos = vmwarevif.get_vif_info(self._session, [ 704.567254] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 704.567818] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] for vif in network_info: [ 704.567818] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 704.567818] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] return self._sync_wrapper(fn, *args, **kwargs) [ 704.567818] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 704.567818] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] self.wait() [ 704.567818] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 704.567818] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] self[:] = self._gt.wait() [ 704.567818] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 704.567818] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] return self._exit_event.wait() [ 704.567818] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 704.567818] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] result = hub.switch() [ 704.567818] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 704.567818] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] return self.greenlet.switch() [ 704.568577] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 704.568577] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] result = function(*args, **kwargs) [ 704.568577] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 704.568577] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] return func(*args, **kwargs) [ 704.568577] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 704.568577] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] raise e [ 704.568577] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 704.568577] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] nwinfo = self.network_api.allocate_for_instance( [ 704.568577] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 704.568577] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] created_port_ids = self._update_ports_for_instance( [ 704.568577] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 704.568577] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] with excutils.save_and_reraise_exception(): [ 704.568577] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 704.569134] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] self.force_reraise() [ 704.569134] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 704.569134] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] raise self.value [ 704.569134] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 704.569134] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] updated_port = self._update_port( [ 704.569134] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 704.569134] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] _ensure_no_port_binding_failure(port) [ 704.569134] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 704.569134] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] raise exception.PortBindingFailed(port_id=port['id']) [ 704.569134] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] nova.exception.PortBindingFailed: Binding failed for port 956ced4a-8d22-4660-b90c-a7a197dd8adb, please check neutron logs for more information. [ 704.569134] env[62460]: ERROR nova.compute.manager [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] [ 704.569774] env[62460]: DEBUG nova.compute.utils [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] Binding failed for port 956ced4a-8d22-4660-b90c-a7a197dd8adb, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 704.569774] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.854s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 704.570735] env[62460]: INFO nova.compute.claims [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 704.573354] env[62460]: DEBUG nova.compute.manager [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] Build of instance 02cb219f-989f-45b5-a9ba-d0b5a521cfa2 was re-scheduled: Binding failed for port 956ced4a-8d22-4660-b90c-a7a197dd8adb, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 704.573787] env[62460]: DEBUG nova.compute.manager [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 704.574055] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Acquiring lock "refresh_cache-02cb219f-989f-45b5-a9ba-d0b5a521cfa2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 704.574176] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Acquired lock "refresh_cache-02cb219f-989f-45b5-a9ba-d0b5a521cfa2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 704.574342] env[62460]: DEBUG nova.network.neutron [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 704.648599] env[62460]: DEBUG nova.network.neutron [req-b9c40289-7f61-4cc3-828f-f54eb0cc2396 req-f6771c48-2ded-4d40-915f-47c574769f60 service nova] [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 704.746040] env[62460]: DEBUG nova.network.neutron [req-b9c40289-7f61-4cc3-828f-f54eb0cc2396 req-f6771c48-2ded-4d40-915f-47c574769f60 service nova] [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.816047] env[62460]: INFO nova.compute.manager [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] [instance: 252df6ad-e29d-4596-bc99-4aae144bbcc5] Took 1.02 seconds to deallocate network for instance. [ 705.042576] env[62460]: INFO nova.compute.manager [-] [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] Took 1.02 seconds to deallocate network for instance. [ 705.044820] env[62460]: DEBUG nova.compute.claims [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 705.044993] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 705.094795] env[62460]: DEBUG nova.network.neutron [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 705.168790] env[62460]: DEBUG nova.network.neutron [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 705.249599] env[62460]: DEBUG oslo_concurrency.lockutils [req-b9c40289-7f61-4cc3-828f-f54eb0cc2396 req-f6771c48-2ded-4d40-915f-47c574769f60 service nova] Releasing lock "refresh_cache-2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 705.249599] env[62460]: DEBUG nova.compute.manager [req-b9c40289-7f61-4cc3-828f-f54eb0cc2396 req-f6771c48-2ded-4d40-915f-47c574769f60 service nova] [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] Received event network-vif-deleted-0c0c5547-5634-45e8-b02c-5c73c6b96852 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 705.670735] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Releasing lock "refresh_cache-02cb219f-989f-45b5-a9ba-d0b5a521cfa2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 705.670996] env[62460]: DEBUG nova.compute.manager [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 705.671158] env[62460]: DEBUG nova.compute.manager [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 705.671329] env[62460]: DEBUG nova.network.neutron [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 705.685605] env[62460]: DEBUG nova.network.neutron [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 705.846065] env[62460]: INFO nova.scheduler.client.report [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] Deleted allocations for instance 252df6ad-e29d-4596-bc99-4aae144bbcc5 [ 705.910450] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64640353-9bd3-46a5-8e57-4f631aa876e8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.918051] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56910f69-488e-4b01-a4c3-960d4f0f27e0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.948500] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-233f9b56-c2c9-4103-920b-b0f183210d94 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.955977] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a4b09b6-f478-4dd0-94fc-743abfcc931f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.968633] env[62460]: DEBUG nova.compute.provider_tree [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 706.190607] env[62460]: DEBUG nova.network.neutron [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 706.355506] env[62460]: DEBUG oslo_concurrency.lockutils [None req-fb15e8b4-1732-4180-abac-6815f7c8c117 tempest-ServerDiagnosticsNegativeTest-1357052768 tempest-ServerDiagnosticsNegativeTest-1357052768-project-member] Lock "252df6ad-e29d-4596-bc99-4aae144bbcc5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 119.190s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 706.473137] env[62460]: DEBUG nova.scheduler.client.report [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 706.693945] env[62460]: INFO nova.compute.manager [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] [instance: 02cb219f-989f-45b5-a9ba-d0b5a521cfa2] Took 1.02 seconds to deallocate network for instance. [ 706.858045] env[62460]: DEBUG nova.compute.manager [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 706.978121] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.408s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 706.978121] env[62460]: DEBUG nova.compute.manager [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 706.983032] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.838s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 707.255063] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 707.255253] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 707.255422] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Starting heal instance info cache {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 707.255545] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Rebuilding the list of instances to heal {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 707.381626] env[62460]: DEBUG oslo_concurrency.lockutils [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 707.488019] env[62460]: DEBUG nova.compute.utils [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 707.492580] env[62460]: DEBUG nova.compute.manager [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 707.492665] env[62460]: DEBUG nova.network.neutron [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 707.575418] env[62460]: DEBUG nova.policy [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8a3ce06e49f64f5c91887c28b4b635cb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '053598069417482780a7387842dd71e0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 707.733123] env[62460]: INFO nova.scheduler.client.report [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Deleted allocations for instance 02cb219f-989f-45b5-a9ba-d0b5a521cfa2 [ 707.760922] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 724a8083-2bde-483e-bd5a-a928def284bf] Skipping network cache update for instance because it is Building. {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 707.761170] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] Skipping network cache update for instance because it is Building. {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 707.761312] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] Skipping network cache update for instance because it is Building. {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 707.761432] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] Skipping network cache update for instance because it is Building. {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 707.761559] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] Skipping network cache update for instance because it is Building. {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 707.761680] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] Skipping network cache update for instance because it is Building. {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 707.761800] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] Skipping network cache update for instance because it is Building. {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 707.761922] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Didn't find any instances for network info cache update. {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 707.762356] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 707.762539] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 707.762691] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 707.762842] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 707.762980] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 707.764054] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 707.764054] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62460) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 707.764054] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 707.863995] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50525e41-fa7d-4711-afed-7b3713c14aa8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.872113] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80fc10e7-f099-4b52-b31d-4fed43790f06 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.902901] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfe5c40f-ffbf-49a0-b830-143701c7f67e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.910630] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8b8abfe-a0a4-42ca-a8ee-baf4ba701770 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.923892] env[62460]: DEBUG nova.compute.provider_tree [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 707.993735] env[62460]: DEBUG nova.compute.manager [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 708.087279] env[62460]: DEBUG nova.network.neutron [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] Successfully created port: 2d5fe2cf-85d6-4ff8-aa4e-a8f56a2c2971 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 708.245463] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4aa4957c-e6d6-4831-af7e-5cf4bbc9b9cb tempest-ServersTestMultiNic-798367234 tempest-ServersTestMultiNic-798367234-project-member] Lock "02cb219f-989f-45b5-a9ba-d0b5a521cfa2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 120.726s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 708.267653] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 708.427510] env[62460]: DEBUG nova.scheduler.client.report [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 708.748750] env[62460]: DEBUG nova.compute.manager [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 708.934043] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.951s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 708.934687] env[62460]: ERROR nova.compute.manager [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] [instance: 578627be-e695-4953-8d0e-9763d12b9a28] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port fa6e34fb-aa15-4b5a-909a-4ed21e7e6fbd, please check neutron logs for more information. [ 708.934687] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] Traceback (most recent call last): [ 708.934687] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 708.934687] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] self.driver.spawn(context, instance, image_meta, [ 708.934687] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 708.934687] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] self._vmops.spawn(context, instance, image_meta, injected_files, [ 708.934687] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 708.934687] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] vm_ref = self.build_virtual_machine(instance, [ 708.934687] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 708.934687] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] vif_infos = vmwarevif.get_vif_info(self._session, [ 708.934687] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 708.935210] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] for vif in network_info: [ 708.935210] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 708.935210] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] return self._sync_wrapper(fn, *args, **kwargs) [ 708.935210] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 708.935210] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] self.wait() [ 708.935210] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 708.935210] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] self[:] = self._gt.wait() [ 708.935210] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 708.935210] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] return self._exit_event.wait() [ 708.935210] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 708.935210] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] current.throw(*self._exc) [ 708.935210] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 708.935210] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] result = function(*args, **kwargs) [ 708.936206] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 708.936206] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] return func(*args, **kwargs) [ 708.936206] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 708.936206] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] raise e [ 708.936206] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 708.936206] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] nwinfo = self.network_api.allocate_for_instance( [ 708.936206] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 708.936206] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] created_port_ids = self._update_ports_for_instance( [ 708.936206] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 708.936206] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] with excutils.save_and_reraise_exception(): [ 708.936206] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 708.936206] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] self.force_reraise() [ 708.936206] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 708.936620] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] raise self.value [ 708.936620] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 708.936620] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] updated_port = self._update_port( [ 708.936620] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 708.936620] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] _ensure_no_port_binding_failure(port) [ 708.936620] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 708.936620] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] raise exception.PortBindingFailed(port_id=port['id']) [ 708.936620] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] nova.exception.PortBindingFailed: Binding failed for port fa6e34fb-aa15-4b5a-909a-4ed21e7e6fbd, please check neutron logs for more information. [ 708.936620] env[62460]: ERROR nova.compute.manager [instance: 578627be-e695-4953-8d0e-9763d12b9a28] [ 708.936620] env[62460]: DEBUG nova.compute.utils [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] [instance: 578627be-e695-4953-8d0e-9763d12b9a28] Binding failed for port fa6e34fb-aa15-4b5a-909a-4ed21e7e6fbd, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 708.938963] env[62460]: DEBUG oslo_concurrency.lockutils [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.107s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 708.946138] env[62460]: DEBUG nova.compute.manager [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] [instance: 578627be-e695-4953-8d0e-9763d12b9a28] Build of instance 578627be-e695-4953-8d0e-9763d12b9a28 was re-scheduled: Binding failed for port fa6e34fb-aa15-4b5a-909a-4ed21e7e6fbd, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 708.946138] env[62460]: DEBUG nova.compute.manager [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] [instance: 578627be-e695-4953-8d0e-9763d12b9a28] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 708.946138] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] Acquiring lock "refresh_cache-578627be-e695-4953-8d0e-9763d12b9a28" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 708.946138] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] Acquired lock "refresh_cache-578627be-e695-4953-8d0e-9763d12b9a28" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 708.946375] env[62460]: DEBUG nova.network.neutron [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] [instance: 578627be-e695-4953-8d0e-9763d12b9a28] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 708.952220] env[62460]: DEBUG nova.compute.manager [req-e895a4d5-5b9d-4e53-91cd-1d939a20a8d6 req-e196e1d4-3d78-492b-bce0-45ca8c17f812 service nova] [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] Received event network-changed-2d5fe2cf-85d6-4ff8-aa4e-a8f56a2c2971 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 708.953140] env[62460]: DEBUG nova.compute.manager [req-e895a4d5-5b9d-4e53-91cd-1d939a20a8d6 req-e196e1d4-3d78-492b-bce0-45ca8c17f812 service nova] [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] Refreshing instance network info cache due to event network-changed-2d5fe2cf-85d6-4ff8-aa4e-a8f56a2c2971. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 708.953392] env[62460]: DEBUG oslo_concurrency.lockutils [req-e895a4d5-5b9d-4e53-91cd-1d939a20a8d6 req-e196e1d4-3d78-492b-bce0-45ca8c17f812 service nova] Acquiring lock "refresh_cache-c37f1b55-7a06-4c93-9492-d2bd6d2bfe83" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 708.953531] env[62460]: DEBUG oslo_concurrency.lockutils [req-e895a4d5-5b9d-4e53-91cd-1d939a20a8d6 req-e196e1d4-3d78-492b-bce0-45ca8c17f812 service nova] Acquired lock "refresh_cache-c37f1b55-7a06-4c93-9492-d2bd6d2bfe83" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 708.953715] env[62460]: DEBUG nova.network.neutron [req-e895a4d5-5b9d-4e53-91cd-1d939a20a8d6 req-e196e1d4-3d78-492b-bce0-45ca8c17f812 service nova] [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] Refreshing network info cache for port 2d5fe2cf-85d6-4ff8-aa4e-a8f56a2c2971 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 709.003013] env[62460]: DEBUG nova.compute.manager [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 709.034674] env[62460]: DEBUG nova.virt.hardware [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 709.034866] env[62460]: DEBUG nova.virt.hardware [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 709.035042] env[62460]: DEBUG nova.virt.hardware [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 709.035232] env[62460]: DEBUG nova.virt.hardware [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 709.035382] env[62460]: DEBUG nova.virt.hardware [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 709.035532] env[62460]: DEBUG nova.virt.hardware [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 709.035741] env[62460]: DEBUG nova.virt.hardware [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 709.035903] env[62460]: DEBUG nova.virt.hardware [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 709.036082] env[62460]: DEBUG nova.virt.hardware [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 709.036249] env[62460]: DEBUG nova.virt.hardware [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 709.036423] env[62460]: DEBUG nova.virt.hardware [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 709.037344] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-348758a0-78bc-417b-bf00-c8d89fdd9242 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.045205] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-359ba706-8dae-4f91-a618-b5f763b3ebd2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.247881] env[62460]: ERROR nova.compute.manager [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2d5fe2cf-85d6-4ff8-aa4e-a8f56a2c2971, please check neutron logs for more information. [ 709.247881] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 709.247881] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 709.247881] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 709.247881] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 709.247881] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 709.247881] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 709.247881] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 709.247881] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 709.247881] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 709.247881] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 709.247881] env[62460]: ERROR nova.compute.manager raise self.value [ 709.247881] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 709.247881] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 709.247881] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 709.247881] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 709.248409] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 709.248409] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 709.248409] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2d5fe2cf-85d6-4ff8-aa4e-a8f56a2c2971, please check neutron logs for more information. [ 709.248409] env[62460]: ERROR nova.compute.manager [ 709.248409] env[62460]: Traceback (most recent call last): [ 709.248409] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 709.248409] env[62460]: listener.cb(fileno) [ 709.248409] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 709.248409] env[62460]: result = function(*args, **kwargs) [ 709.248409] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 709.248409] env[62460]: return func(*args, **kwargs) [ 709.248409] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 709.248409] env[62460]: raise e [ 709.248409] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 709.248409] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 709.248409] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 709.248409] env[62460]: created_port_ids = self._update_ports_for_instance( [ 709.248409] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 709.248409] env[62460]: with excutils.save_and_reraise_exception(): [ 709.248409] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 709.248409] env[62460]: self.force_reraise() [ 709.248409] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 709.248409] env[62460]: raise self.value [ 709.248409] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 709.248409] env[62460]: updated_port = self._update_port( [ 709.248409] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 709.248409] env[62460]: _ensure_no_port_binding_failure(port) [ 709.248409] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 709.248409] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 709.249279] env[62460]: nova.exception.PortBindingFailed: Binding failed for port 2d5fe2cf-85d6-4ff8-aa4e-a8f56a2c2971, please check neutron logs for more information. [ 709.249279] env[62460]: Removing descriptor: 18 [ 709.249279] env[62460]: ERROR nova.compute.manager [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2d5fe2cf-85d6-4ff8-aa4e-a8f56a2c2971, please check neutron logs for more information. [ 709.249279] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] Traceback (most recent call last): [ 709.249279] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 709.249279] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] yield resources [ 709.249279] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 709.249279] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] self.driver.spawn(context, instance, image_meta, [ 709.249279] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 709.249279] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] self._vmops.spawn(context, instance, image_meta, injected_files, [ 709.249279] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 709.249279] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] vm_ref = self.build_virtual_machine(instance, [ 709.249656] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 709.249656] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] vif_infos = vmwarevif.get_vif_info(self._session, [ 709.249656] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 709.249656] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] for vif in network_info: [ 709.249656] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 709.249656] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] return self._sync_wrapper(fn, *args, **kwargs) [ 709.249656] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 709.249656] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] self.wait() [ 709.249656] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 709.249656] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] self[:] = self._gt.wait() [ 709.249656] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 709.249656] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] return self._exit_event.wait() [ 709.249656] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 709.250076] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] result = hub.switch() [ 709.250076] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 709.250076] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] return self.greenlet.switch() [ 709.250076] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 709.250076] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] result = function(*args, **kwargs) [ 709.250076] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 709.250076] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] return func(*args, **kwargs) [ 709.250076] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 709.250076] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] raise e [ 709.250076] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 709.250076] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] nwinfo = self.network_api.allocate_for_instance( [ 709.250076] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 709.250076] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] created_port_ids = self._update_ports_for_instance( [ 709.250496] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 709.250496] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] with excutils.save_and_reraise_exception(): [ 709.250496] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 709.250496] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] self.force_reraise() [ 709.250496] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 709.250496] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] raise self.value [ 709.250496] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 709.250496] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] updated_port = self._update_port( [ 709.250496] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 709.250496] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] _ensure_no_port_binding_failure(port) [ 709.250496] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 709.250496] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] raise exception.PortBindingFailed(port_id=port['id']) [ 709.250867] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] nova.exception.PortBindingFailed: Binding failed for port 2d5fe2cf-85d6-4ff8-aa4e-a8f56a2c2971, please check neutron logs for more information. [ 709.250867] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] [ 709.250867] env[62460]: INFO nova.compute.manager [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] Terminating instance [ 709.253583] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] Acquiring lock "refresh_cache-c37f1b55-7a06-4c93-9492-d2bd6d2bfe83" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 709.269708] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 709.471429] env[62460]: DEBUG nova.network.neutron [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] [instance: 578627be-e695-4953-8d0e-9763d12b9a28] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 709.489989] env[62460]: DEBUG nova.network.neutron [req-e895a4d5-5b9d-4e53-91cd-1d939a20a8d6 req-e196e1d4-3d78-492b-bce0-45ca8c17f812 service nova] [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 709.614632] env[62460]: DEBUG nova.network.neutron [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] [instance: 578627be-e695-4953-8d0e-9763d12b9a28] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.752670] env[62460]: DEBUG nova.network.neutron [req-e895a4d5-5b9d-4e53-91cd-1d939a20a8d6 req-e196e1d4-3d78-492b-bce0-45ca8c17f812 service nova] [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.813169] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90714ddf-9775-4151-99fe-ada2270d10da {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.821656] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12b1f085-9737-4d4c-8cf5-11b27c73434c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.853016] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-677641f5-aec5-4f6c-bace-b4a1f3c0a39a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.861696] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f746061-55f5-44d0-a928-45de63be3b59 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.879022] env[62460]: DEBUG nova.compute.provider_tree [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 710.118938] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] Releasing lock "refresh_cache-578627be-e695-4953-8d0e-9763d12b9a28" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 710.119328] env[62460]: DEBUG nova.compute.manager [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 710.119418] env[62460]: DEBUG nova.compute.manager [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] [instance: 578627be-e695-4953-8d0e-9763d12b9a28] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 710.119590] env[62460]: DEBUG nova.network.neutron [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] [instance: 578627be-e695-4953-8d0e-9763d12b9a28] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 710.146528] env[62460]: DEBUG nova.network.neutron [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] [instance: 578627be-e695-4953-8d0e-9763d12b9a28] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 710.259674] env[62460]: DEBUG oslo_concurrency.lockutils [req-e895a4d5-5b9d-4e53-91cd-1d939a20a8d6 req-e196e1d4-3d78-492b-bce0-45ca8c17f812 service nova] Releasing lock "refresh_cache-c37f1b55-7a06-4c93-9492-d2bd6d2bfe83" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 710.260126] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] Acquired lock "refresh_cache-c37f1b55-7a06-4c93-9492-d2bd6d2bfe83" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.260318] env[62460]: DEBUG nova.network.neutron [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 710.380718] env[62460]: DEBUG nova.scheduler.client.report [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 710.659318] env[62460]: DEBUG nova.network.neutron [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] [instance: 578627be-e695-4953-8d0e-9763d12b9a28] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 710.791917] env[62460]: DEBUG nova.network.neutron [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 710.885147] env[62460]: DEBUG nova.network.neutron [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 710.888568] env[62460]: DEBUG oslo_concurrency.lockutils [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.951s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 710.889132] env[62460]: ERROR nova.compute.manager [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] [instance: 724a8083-2bde-483e-bd5a-a928def284bf] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7464bd25-e8a6-4cf7-8efd-1f64df719868, please check neutron logs for more information. [ 710.889132] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] Traceback (most recent call last): [ 710.889132] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 710.889132] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] self.driver.spawn(context, instance, image_meta, [ 710.889132] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 710.889132] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 710.889132] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 710.889132] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] vm_ref = self.build_virtual_machine(instance, [ 710.889132] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 710.889132] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] vif_infos = vmwarevif.get_vif_info(self._session, [ 710.889132] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 710.891410] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] for vif in network_info: [ 710.891410] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 710.891410] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] return self._sync_wrapper(fn, *args, **kwargs) [ 710.891410] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 710.891410] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] self.wait() [ 710.891410] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 710.891410] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] self[:] = self._gt.wait() [ 710.891410] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 710.891410] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] return self._exit_event.wait() [ 710.891410] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 710.891410] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] current.throw(*self._exc) [ 710.891410] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 710.891410] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] result = function(*args, **kwargs) [ 710.891896] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 710.891896] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] return func(*args, **kwargs) [ 710.891896] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 710.891896] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] raise e [ 710.891896] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 710.891896] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] nwinfo = self.network_api.allocate_for_instance( [ 710.891896] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 710.891896] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] created_port_ids = self._update_ports_for_instance( [ 710.891896] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 710.891896] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] with excutils.save_and_reraise_exception(): [ 710.891896] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 710.891896] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] self.force_reraise() [ 710.891896] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 710.892245] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] raise self.value [ 710.892245] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 710.892245] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] updated_port = self._update_port( [ 710.892245] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 710.892245] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] _ensure_no_port_binding_failure(port) [ 710.892245] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 710.892245] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] raise exception.PortBindingFailed(port_id=port['id']) [ 710.892245] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] nova.exception.PortBindingFailed: Binding failed for port 7464bd25-e8a6-4cf7-8efd-1f64df719868, please check neutron logs for more information. [ 710.892245] env[62460]: ERROR nova.compute.manager [instance: 724a8083-2bde-483e-bd5a-a928def284bf] [ 710.892245] env[62460]: DEBUG nova.compute.utils [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] [instance: 724a8083-2bde-483e-bd5a-a928def284bf] Binding failed for port 7464bd25-e8a6-4cf7-8efd-1f64df719868, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 710.892512] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.609s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 710.895667] env[62460]: DEBUG nova.compute.manager [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] [instance: 724a8083-2bde-483e-bd5a-a928def284bf] Build of instance 724a8083-2bde-483e-bd5a-a928def284bf was re-scheduled: Binding failed for port 7464bd25-e8a6-4cf7-8efd-1f64df719868, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 710.896139] env[62460]: DEBUG nova.compute.manager [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] [instance: 724a8083-2bde-483e-bd5a-a928def284bf] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 710.896372] env[62460]: DEBUG oslo_concurrency.lockutils [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] Acquiring lock "refresh_cache-724a8083-2bde-483e-bd5a-a928def284bf" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 710.896738] env[62460]: DEBUG oslo_concurrency.lockutils [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] Acquired lock "refresh_cache-724a8083-2bde-483e-bd5a-a928def284bf" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.896963] env[62460]: DEBUG nova.network.neutron [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] [instance: 724a8083-2bde-483e-bd5a-a928def284bf] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 711.036248] env[62460]: DEBUG nova.compute.manager [req-acd2d951-c98e-4862-82a8-0b00ca276a9f req-caaec838-9877-4296-9d2d-4d96c6479094 service nova] [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] Received event network-vif-deleted-2d5fe2cf-85d6-4ff8-aa4e-a8f56a2c2971 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 711.162378] env[62460]: INFO nova.compute.manager [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] [instance: 578627be-e695-4953-8d0e-9763d12b9a28] Took 1.04 seconds to deallocate network for instance. [ 711.386938] env[62460]: DEBUG oslo_concurrency.lockutils [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Acquiring lock "db09d1f5-88cc-4dc7-9a7b-5d53d09567fa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 711.387187] env[62460]: DEBUG oslo_concurrency.lockutils [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Lock "db09d1f5-88cc-4dc7-9a7b-5d53d09567fa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 711.387546] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] Releasing lock "refresh_cache-c37f1b55-7a06-4c93-9492-d2bd6d2bfe83" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 711.387996] env[62460]: DEBUG nova.compute.manager [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 711.388137] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 711.388413] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d6794283-d5a4-44eb-9546-5fb81879a0f7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.400094] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d575cb00-cb88-44c9-ad13-cd4db6ec32ec {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.426478] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c37f1b55-7a06-4c93-9492-d2bd6d2bfe83 could not be found. [ 711.426478] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 711.426478] env[62460]: INFO nova.compute.manager [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] Took 0.04 seconds to destroy the instance on the hypervisor. [ 711.426712] env[62460]: DEBUG oslo.service.loopingcall [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 711.428584] env[62460]: DEBUG nova.compute.manager [-] [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 711.428584] env[62460]: DEBUG nova.network.neutron [-] [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 711.431228] env[62460]: DEBUG nova.network.neutron [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] [instance: 724a8083-2bde-483e-bd5a-a928def284bf] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 711.444140] env[62460]: DEBUG nova.network.neutron [-] [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 711.551844] env[62460]: DEBUG nova.network.neutron [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] [instance: 724a8083-2bde-483e-bd5a-a928def284bf] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 711.732208] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15269cf9-cfd4-46d3-a13a-faff81840f38 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.740180] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64acf6e8-640d-43a9-bee2-c1deb063eebf {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.773877] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52708031-621b-4141-b7df-9815bbb24509 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.781418] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3375315c-bd5e-4a45-b4c3-1616ba9bec16 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.795814] env[62460]: DEBUG nova.compute.provider_tree [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 711.945167] env[62460]: DEBUG nova.network.neutron [-] [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.057071] env[62460]: DEBUG oslo_concurrency.lockutils [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] Releasing lock "refresh_cache-724a8083-2bde-483e-bd5a-a928def284bf" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 712.057071] env[62460]: DEBUG nova.compute.manager [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 712.057071] env[62460]: DEBUG nova.compute.manager [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] [instance: 724a8083-2bde-483e-bd5a-a928def284bf] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 712.057071] env[62460]: DEBUG nova.network.neutron [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] [instance: 724a8083-2bde-483e-bd5a-a928def284bf] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 712.077050] env[62460]: DEBUG nova.network.neutron [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] [instance: 724a8083-2bde-483e-bd5a-a928def284bf] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 712.207769] env[62460]: INFO nova.scheduler.client.report [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] Deleted allocations for instance 578627be-e695-4953-8d0e-9763d12b9a28 [ 712.302949] env[62460]: DEBUG nova.scheduler.client.report [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 712.448464] env[62460]: INFO nova.compute.manager [-] [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] Took 1.02 seconds to deallocate network for instance. [ 712.449893] env[62460]: DEBUG nova.compute.claims [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 712.450190] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 712.580261] env[62460]: DEBUG nova.network.neutron [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] [instance: 724a8083-2bde-483e-bd5a-a928def284bf] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.727869] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0ce0ba53-c8b7-47db-973a-e5bf9eb4971b tempest-ServerActionsTestOtherB-198155639 tempest-ServerActionsTestOtherB-198155639-project-member] Lock "578627be-e695-4953-8d0e-9763d12b9a28" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 124.315s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 712.808131] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.917s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 712.808810] env[62460]: ERROR nova.compute.manager [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c4831938-0e88-4552-9ca7-091fbd50dd0f, please check neutron logs for more information. [ 712.808810] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] Traceback (most recent call last): [ 712.808810] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 712.808810] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] self.driver.spawn(context, instance, image_meta, [ 712.808810] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 712.808810] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] self._vmops.spawn(context, instance, image_meta, injected_files, [ 712.808810] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 712.808810] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] vm_ref = self.build_virtual_machine(instance, [ 712.808810] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 712.808810] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] vif_infos = vmwarevif.get_vif_info(self._session, [ 712.808810] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 712.809253] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] for vif in network_info: [ 712.809253] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 712.809253] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] return self._sync_wrapper(fn, *args, **kwargs) [ 712.809253] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 712.809253] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] self.wait() [ 712.809253] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 712.809253] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] self[:] = self._gt.wait() [ 712.809253] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 712.809253] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] return self._exit_event.wait() [ 712.809253] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 712.809253] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] current.throw(*self._exc) [ 712.809253] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 712.809253] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] result = function(*args, **kwargs) [ 712.809643] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 712.809643] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] return func(*args, **kwargs) [ 712.809643] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 712.809643] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] raise e [ 712.809643] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 712.809643] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] nwinfo = self.network_api.allocate_for_instance( [ 712.809643] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 712.809643] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] created_port_ids = self._update_ports_for_instance( [ 712.809643] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 712.809643] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] with excutils.save_and_reraise_exception(): [ 712.809643] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 712.809643] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] self.force_reraise() [ 712.809643] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 712.810038] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] raise self.value [ 712.810038] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 712.810038] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] updated_port = self._update_port( [ 712.810038] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 712.810038] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] _ensure_no_port_binding_failure(port) [ 712.810038] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 712.810038] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] raise exception.PortBindingFailed(port_id=port['id']) [ 712.810038] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] nova.exception.PortBindingFailed: Binding failed for port c4831938-0e88-4552-9ca7-091fbd50dd0f, please check neutron logs for more information. [ 712.810038] env[62460]: ERROR nova.compute.manager [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] [ 712.810038] env[62460]: DEBUG nova.compute.utils [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] Binding failed for port c4831938-0e88-4552-9ca7-091fbd50dd0f, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 712.811193] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.109s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 712.814216] env[62460]: DEBUG nova.compute.manager [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] Build of instance 0d29a4ce-6689-4d0c-8144-e05853eb3a60 was re-scheduled: Binding failed for port c4831938-0e88-4552-9ca7-091fbd50dd0f, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 712.814901] env[62460]: DEBUG nova.compute.manager [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 712.815133] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Acquiring lock "refresh_cache-0d29a4ce-6689-4d0c-8144-e05853eb3a60" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.815331] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Acquired lock "refresh_cache-0d29a4ce-6689-4d0c-8144-e05853eb3a60" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 712.815521] env[62460]: DEBUG nova.network.neutron [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 713.083327] env[62460]: INFO nova.compute.manager [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] [instance: 724a8083-2bde-483e-bd5a-a928def284bf] Took 1.03 seconds to deallocate network for instance. [ 713.231095] env[62460]: DEBUG nova.compute.manager [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 713.351292] env[62460]: DEBUG nova.network.neutron [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 713.527796] env[62460]: DEBUG nova.network.neutron [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.712988] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b364781-f3ae-407d-a3ab-6e29c6cfc28a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.721595] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c9ca0dc-b8a0-43b4-aabe-915377d4c142 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.766046] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b663faf9-6031-4739-9f9e-6a744a384eea {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.775750] env[62460]: DEBUG oslo_concurrency.lockutils [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Acquiring lock "72e91f8e-0619-464c-b9bc-d6a14be42cb8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 713.775750] env[62460]: DEBUG oslo_concurrency.lockutils [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Lock "72e91f8e-0619-464c-b9bc-d6a14be42cb8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 713.780202] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-147666e5-0618-4922-9b4f-14322095af2c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.786196] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 713.798424] env[62460]: DEBUG nova.compute.provider_tree [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 714.033725] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Releasing lock "refresh_cache-0d29a4ce-6689-4d0c-8144-e05853eb3a60" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 714.033961] env[62460]: DEBUG nova.compute.manager [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 714.034213] env[62460]: DEBUG nova.compute.manager [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 714.034549] env[62460]: DEBUG nova.network.neutron [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 714.050252] env[62460]: DEBUG nova.network.neutron [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 714.125542] env[62460]: DEBUG oslo_concurrency.lockutils [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Acquiring lock "3ada3516-3147-4566-a46a-1cb29cf880d0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 714.125900] env[62460]: DEBUG oslo_concurrency.lockutils [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Lock "3ada3516-3147-4566-a46a-1cb29cf880d0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 714.128941] env[62460]: INFO nova.scheduler.client.report [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] Deleted allocations for instance 724a8083-2bde-483e-bd5a-a928def284bf [ 714.302128] env[62460]: DEBUG nova.scheduler.client.report [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 714.553122] env[62460]: DEBUG nova.network.neutron [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.620802] env[62460]: DEBUG oslo_concurrency.lockutils [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Acquiring lock "fde12685-d3b9-46a0-8931-25b904d4f21e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 714.621270] env[62460]: DEBUG oslo_concurrency.lockutils [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Lock "fde12685-d3b9-46a0-8931-25b904d4f21e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 714.639920] env[62460]: DEBUG oslo_concurrency.lockutils [None req-41a0834d-09f7-4dc5-81ea-507d0f468fc2 tempest-ServersTestJSON-782287380 tempest-ServersTestJSON-782287380-project-member] Lock "724a8083-2bde-483e-bd5a-a928def284bf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 125.719s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 714.813605] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.002s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 714.814368] env[62460]: ERROR nova.compute.manager [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8e4514d2-11c6-4bd9-a27f-0b171dad4c9d, please check neutron logs for more information. [ 714.814368] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] Traceback (most recent call last): [ 714.814368] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 714.814368] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] self.driver.spawn(context, instance, image_meta, [ 714.814368] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 714.814368] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] self._vmops.spawn(context, instance, image_meta, injected_files, [ 714.814368] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 714.814368] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] vm_ref = self.build_virtual_machine(instance, [ 714.814368] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 714.814368] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] vif_infos = vmwarevif.get_vif_info(self._session, [ 714.814368] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 714.814741] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] for vif in network_info: [ 714.814741] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 714.814741] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] return self._sync_wrapper(fn, *args, **kwargs) [ 714.814741] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 714.814741] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] self.wait() [ 714.814741] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 714.814741] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] self[:] = self._gt.wait() [ 714.814741] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 714.814741] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] return self._exit_event.wait() [ 714.814741] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 714.814741] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] current.throw(*self._exc) [ 714.814741] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 714.814741] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] result = function(*args, **kwargs) [ 714.815175] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 714.815175] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] return func(*args, **kwargs) [ 714.815175] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 714.815175] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] raise e [ 714.815175] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 714.815175] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] nwinfo = self.network_api.allocate_for_instance( [ 714.815175] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 714.815175] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] created_port_ids = self._update_ports_for_instance( [ 714.815175] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 714.815175] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] with excutils.save_and_reraise_exception(): [ 714.815175] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 714.815175] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] self.force_reraise() [ 714.815175] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 714.815593] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] raise self.value [ 714.815593] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 714.815593] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] updated_port = self._update_port( [ 714.815593] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 714.815593] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] _ensure_no_port_binding_failure(port) [ 714.815593] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 714.815593] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] raise exception.PortBindingFailed(port_id=port['id']) [ 714.815593] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] nova.exception.PortBindingFailed: Binding failed for port 8e4514d2-11c6-4bd9-a27f-0b171dad4c9d, please check neutron logs for more information. [ 714.815593] env[62460]: ERROR nova.compute.manager [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] [ 714.815593] env[62460]: DEBUG nova.compute.utils [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] Binding failed for port 8e4514d2-11c6-4bd9-a27f-0b171dad4c9d, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 714.816712] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.680s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 714.819642] env[62460]: DEBUG nova.compute.manager [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] Build of instance 02b7a3c3-bdfd-47d9-acf7-afd7725cea64 was re-scheduled: Binding failed for port 8e4514d2-11c6-4bd9-a27f-0b171dad4c9d, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 714.820089] env[62460]: DEBUG nova.compute.manager [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 714.820313] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Acquiring lock "refresh_cache-02b7a3c3-bdfd-47d9-acf7-afd7725cea64" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 714.820476] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Acquired lock "refresh_cache-02b7a3c3-bdfd-47d9-acf7-afd7725cea64" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.820661] env[62460]: DEBUG nova.network.neutron [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 715.055874] env[62460]: INFO nova.compute.manager [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 0d29a4ce-6689-4d0c-8144-e05853eb3a60] Took 1.02 seconds to deallocate network for instance. [ 715.143618] env[62460]: DEBUG nova.compute.manager [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 715.353746] env[62460]: DEBUG nova.network.neutron [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 715.496549] env[62460]: DEBUG nova.network.neutron [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.668518] env[62460]: DEBUG oslo_concurrency.lockutils [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 715.714065] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bd5c6ef-d48d-4166-951a-8e27f5e03db9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.728174] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06a26f2c-8b3d-4fb8-b707-744f60e8e501 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.760928] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c686e6dd-86aa-4898-828d-0119d124ef4b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.769123] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cbdb11d-700f-4cea-93ae-191986c7cb5c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.782648] env[62460]: DEBUG nova.compute.provider_tree [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 715.999549] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Releasing lock "refresh_cache-02b7a3c3-bdfd-47d9-acf7-afd7725cea64" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 716.000035] env[62460]: DEBUG nova.compute.manager [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 716.002440] env[62460]: DEBUG nova.compute.manager [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 716.002440] env[62460]: DEBUG nova.network.neutron [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 716.018520] env[62460]: DEBUG nova.network.neutron [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 716.091305] env[62460]: INFO nova.scheduler.client.report [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Deleted allocations for instance 0d29a4ce-6689-4d0c-8144-e05853eb3a60 [ 716.286315] env[62460]: DEBUG nova.scheduler.client.report [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 716.520598] env[62460]: DEBUG nova.network.neutron [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 716.600882] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5036a5f3-4a94-4af3-b618-bfec19a37846 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Lock "0d29a4ce-6689-4d0c-8144-e05853eb3a60" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 126.240s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 716.792141] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.975s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 716.792773] env[62460]: ERROR nova.compute.manager [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port bf96148d-c685-42c5-a715-fec0fd939e37, please check neutron logs for more information. [ 716.792773] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] Traceback (most recent call last): [ 716.792773] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 716.792773] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] self.driver.spawn(context, instance, image_meta, [ 716.792773] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 716.792773] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 716.792773] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 716.792773] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] vm_ref = self.build_virtual_machine(instance, [ 716.792773] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 716.792773] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] vif_infos = vmwarevif.get_vif_info(self._session, [ 716.792773] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 716.793204] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] for vif in network_info: [ 716.793204] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 716.793204] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] return self._sync_wrapper(fn, *args, **kwargs) [ 716.793204] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 716.793204] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] self.wait() [ 716.793204] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 716.793204] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] self[:] = self._gt.wait() [ 716.793204] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 716.793204] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] return self._exit_event.wait() [ 716.793204] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 716.793204] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] current.throw(*self._exc) [ 716.793204] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 716.793204] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] result = function(*args, **kwargs) [ 716.793702] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 716.793702] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] return func(*args, **kwargs) [ 716.793702] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 716.793702] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] raise e [ 716.793702] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 716.793702] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] nwinfo = self.network_api.allocate_for_instance( [ 716.793702] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 716.793702] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] created_port_ids = self._update_ports_for_instance( [ 716.793702] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 716.793702] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] with excutils.save_and_reraise_exception(): [ 716.793702] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 716.793702] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] self.force_reraise() [ 716.793702] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 716.795469] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] raise self.value [ 716.795469] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 716.795469] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] updated_port = self._update_port( [ 716.795469] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 716.795469] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] _ensure_no_port_binding_failure(port) [ 716.795469] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 716.795469] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] raise exception.PortBindingFailed(port_id=port['id']) [ 716.795469] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] nova.exception.PortBindingFailed: Binding failed for port bf96148d-c685-42c5-a715-fec0fd939e37, please check neutron logs for more information. [ 716.795469] env[62460]: ERROR nova.compute.manager [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] [ 716.795469] env[62460]: DEBUG nova.compute.utils [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] Binding failed for port bf96148d-c685-42c5-a715-fec0fd939e37, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 716.795839] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.446s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 716.798719] env[62460]: DEBUG nova.compute.manager [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] Build of instance 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4 was re-scheduled: Binding failed for port bf96148d-c685-42c5-a715-fec0fd939e37, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 716.799104] env[62460]: DEBUG nova.compute.manager [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 716.799659] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] Acquiring lock "refresh_cache-47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 716.799902] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] Acquired lock "refresh_cache-47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.800584] env[62460]: DEBUG nova.network.neutron [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 717.024930] env[62460]: INFO nova.compute.manager [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] [instance: 02b7a3c3-bdfd-47d9-acf7-afd7725cea64] Took 1.02 seconds to deallocate network for instance. [ 717.105478] env[62460]: DEBUG nova.compute.manager [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 4ed90d16-81a6-4dbd-8936-0e137151171f] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 717.327991] env[62460]: DEBUG nova.network.neutron [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 717.431694] env[62460]: DEBUG nova.network.neutron [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 717.633332] env[62460]: DEBUG oslo_concurrency.lockutils [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.650707] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e902b137-948b-48ad-8acb-15099f410ab6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.659061] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3edd0db1-ab36-4942-801d-19f94923a840 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.690293] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4eb20801-0d6f-4a88-987d-4f1508163364 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.697984] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-085e470f-1c0c-42dd-84c3-3f66dcf31991 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.714165] env[62460]: DEBUG nova.compute.provider_tree [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 717.933445] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] Releasing lock "refresh_cache-47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.933556] env[62460]: DEBUG nova.compute.manager [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 717.933742] env[62460]: DEBUG nova.compute.manager [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 717.933920] env[62460]: DEBUG nova.network.neutron [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 717.949591] env[62460]: DEBUG nova.network.neutron [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 718.064052] env[62460]: INFO nova.scheduler.client.report [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Deleted allocations for instance 02b7a3c3-bdfd-47d9-acf7-afd7725cea64 [ 718.217609] env[62460]: DEBUG nova.scheduler.client.report [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 718.452471] env[62460]: DEBUG nova.network.neutron [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.574342] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a5afff24-9989-4ded-ad85-3bd4f89a9c83 tempest-ServerRescueNegativeTestJSON-606921365 tempest-ServerRescueNegativeTestJSON-606921365-project-member] Lock "02b7a3c3-bdfd-47d9-acf7-afd7725cea64" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 126.963s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.725563] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.930s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.726223] env[62460]: ERROR nova.compute.manager [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b3b66865-1822-47a2-a3d1-45a61b73a2cf, please check neutron logs for more information. [ 718.726223] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] Traceback (most recent call last): [ 718.726223] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 718.726223] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] self.driver.spawn(context, instance, image_meta, [ 718.726223] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 718.726223] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] self._vmops.spawn(context, instance, image_meta, injected_files, [ 718.726223] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 718.726223] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] vm_ref = self.build_virtual_machine(instance, [ 718.726223] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 718.726223] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] vif_infos = vmwarevif.get_vif_info(self._session, [ 718.726223] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 718.726649] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] for vif in network_info: [ 718.726649] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 718.726649] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] return self._sync_wrapper(fn, *args, **kwargs) [ 718.726649] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 718.726649] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] self.wait() [ 718.726649] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 718.726649] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] self[:] = self._gt.wait() [ 718.726649] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 718.726649] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] return self._exit_event.wait() [ 718.726649] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 718.726649] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] current.throw(*self._exc) [ 718.726649] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 718.726649] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] result = function(*args, **kwargs) [ 718.727150] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 718.727150] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] return func(*args, **kwargs) [ 718.727150] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 718.727150] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] raise e [ 718.727150] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 718.727150] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] nwinfo = self.network_api.allocate_for_instance( [ 718.727150] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 718.727150] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] created_port_ids = self._update_ports_for_instance( [ 718.727150] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 718.727150] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] with excutils.save_and_reraise_exception(): [ 718.727150] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 718.727150] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] self.force_reraise() [ 718.727150] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 718.727513] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] raise self.value [ 718.727513] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 718.727513] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] updated_port = self._update_port( [ 718.727513] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 718.727513] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] _ensure_no_port_binding_failure(port) [ 718.727513] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 718.727513] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] raise exception.PortBindingFailed(port_id=port['id']) [ 718.727513] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] nova.exception.PortBindingFailed: Binding failed for port b3b66865-1822-47a2-a3d1-45a61b73a2cf, please check neutron logs for more information. [ 718.727513] env[62460]: ERROR nova.compute.manager [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] [ 718.727513] env[62460]: DEBUG nova.compute.utils [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] Binding failed for port b3b66865-1822-47a2-a3d1-45a61b73a2cf, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 718.728137] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.683s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 718.731237] env[62460]: DEBUG nova.compute.manager [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] Build of instance 770b6195-ab31-4df0-84b0-e8382732ec32 was re-scheduled: Binding failed for port b3b66865-1822-47a2-a3d1-45a61b73a2cf, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 718.731667] env[62460]: DEBUG nova.compute.manager [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 718.731957] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] Acquiring lock "refresh_cache-770b6195-ab31-4df0-84b0-e8382732ec32" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 718.732050] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] Acquired lock "refresh_cache-770b6195-ab31-4df0-84b0-e8382732ec32" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.732246] env[62460]: DEBUG nova.network.neutron [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 718.956297] env[62460]: INFO nova.compute.manager [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] [instance: 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4] Took 1.02 seconds to deallocate network for instance. [ 719.077468] env[62460]: DEBUG nova.compute.manager [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 719.264305] env[62460]: DEBUG nova.network.neutron [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 719.437016] env[62460]: DEBUG nova.network.neutron [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.603615] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 719.617429] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b20826a-2c90-4cf1-aef0-dd0e739fd4ce {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.624947] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d073881-eaf7-490a-bb31-a3c02ba007e8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.654395] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6ae6bab-f219-4ac5-9ad5-ac03b40cf566 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.661930] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b86c783-cfd7-4888-8e05-cf278a01440b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.677453] env[62460]: DEBUG nova.compute.provider_tree [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 719.944700] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] Releasing lock "refresh_cache-770b6195-ab31-4df0-84b0-e8382732ec32" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.944961] env[62460]: DEBUG nova.compute.manager [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 719.945172] env[62460]: DEBUG nova.compute.manager [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 719.945346] env[62460]: DEBUG nova.network.neutron [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 719.963848] env[62460]: DEBUG nova.network.neutron [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 719.994148] env[62460]: INFO nova.scheduler.client.report [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] Deleted allocations for instance 47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4 [ 720.181753] env[62460]: DEBUG nova.scheduler.client.report [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 720.467009] env[62460]: DEBUG nova.network.neutron [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.508706] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ded32dcd-5c71-42d9-9444-091ea8a3ec9e tempest-ServersV294TestFqdnHostnames-1041709389 tempest-ServersV294TestFqdnHostnames-1041709389-project-member] Lock "47e44ea5-15f2-4ea3-bda9-ae4820c4b9c4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 124.812s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 720.688932] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.961s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 720.689682] env[62460]: ERROR nova.compute.manager [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0c0c5547-5634-45e8-b02c-5c73c6b96852, please check neutron logs for more information. [ 720.689682] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] Traceback (most recent call last): [ 720.689682] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 720.689682] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] self.driver.spawn(context, instance, image_meta, [ 720.689682] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 720.689682] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] self._vmops.spawn(context, instance, image_meta, injected_files, [ 720.689682] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 720.689682] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] vm_ref = self.build_virtual_machine(instance, [ 720.689682] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 720.689682] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] vif_infos = vmwarevif.get_vif_info(self._session, [ 720.689682] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 720.690022] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] for vif in network_info: [ 720.690022] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 720.690022] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] return self._sync_wrapper(fn, *args, **kwargs) [ 720.690022] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 720.690022] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] self.wait() [ 720.690022] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 720.690022] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] self[:] = self._gt.wait() [ 720.690022] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 720.690022] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] return self._exit_event.wait() [ 720.690022] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 720.690022] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] current.throw(*self._exc) [ 720.690022] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 720.690022] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] result = function(*args, **kwargs) [ 720.690632] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 720.690632] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] return func(*args, **kwargs) [ 720.690632] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 720.690632] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] raise e [ 720.690632] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 720.690632] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] nwinfo = self.network_api.allocate_for_instance( [ 720.690632] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 720.690632] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] created_port_ids = self._update_ports_for_instance( [ 720.690632] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 720.690632] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] with excutils.save_and_reraise_exception(): [ 720.690632] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 720.690632] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] self.force_reraise() [ 720.690632] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 720.691216] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] raise self.value [ 720.691216] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 720.691216] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] updated_port = self._update_port( [ 720.691216] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 720.691216] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] _ensure_no_port_binding_failure(port) [ 720.691216] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 720.691216] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] raise exception.PortBindingFailed(port_id=port['id']) [ 720.691216] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] nova.exception.PortBindingFailed: Binding failed for port 0c0c5547-5634-45e8-b02c-5c73c6b96852, please check neutron logs for more information. [ 720.691216] env[62460]: ERROR nova.compute.manager [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] [ 720.691216] env[62460]: DEBUG nova.compute.utils [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] Binding failed for port 0c0c5547-5634-45e8-b02c-5c73c6b96852, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 720.691738] env[62460]: DEBUG oslo_concurrency.lockutils [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.310s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 720.693188] env[62460]: INFO nova.compute.claims [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 720.695811] env[62460]: DEBUG nova.compute.manager [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] Build of instance 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae was re-scheduled: Binding failed for port 0c0c5547-5634-45e8-b02c-5c73c6b96852, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 720.696270] env[62460]: DEBUG nova.compute.manager [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 720.696500] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] Acquiring lock "refresh_cache-2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 720.696648] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] Acquired lock "refresh_cache-2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.696811] env[62460]: DEBUG nova.network.neutron [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 720.970394] env[62460]: INFO nova.compute.manager [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] [instance: 770b6195-ab31-4df0-84b0-e8382732ec32] Took 1.02 seconds to deallocate network for instance. [ 721.013545] env[62460]: DEBUG nova.compute.manager [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 721.223921] env[62460]: DEBUG nova.network.neutron [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 721.325915] env[62460]: DEBUG nova.network.neutron [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.473335] env[62460]: DEBUG oslo_concurrency.lockutils [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Acquiring lock "13890351-6091-4b4f-8484-1cdd0c8523b0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 721.474665] env[62460]: DEBUG oslo_concurrency.lockutils [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Lock "13890351-6091-4b4f-8484-1cdd0c8523b0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 721.539719] env[62460]: DEBUG oslo_concurrency.lockutils [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 721.832170] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] Releasing lock "refresh_cache-2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 721.832170] env[62460]: DEBUG nova.compute.manager [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 721.832170] env[62460]: DEBUG nova.compute.manager [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 721.832170] env[62460]: DEBUG nova.network.neutron [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 721.869832] env[62460]: DEBUG nova.network.neutron [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 722.024324] env[62460]: INFO nova.scheduler.client.report [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] Deleted allocations for instance 770b6195-ab31-4df0-84b0-e8382732ec32 [ 722.103836] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4338ae25-41e7-4ee7-8208-d4890c07b247 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.115232] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0d2725c-2343-43cd-b04c-10c5f843f946 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.146524] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cf4990c-4aa3-43fe-9616-e681cc862de9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.154200] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-946210eb-48fd-45f7-b1df-7ea863552e1d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.169816] env[62460]: DEBUG nova.compute.provider_tree [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 722.375305] env[62460]: DEBUG nova.network.neutron [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 722.529864] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c8bd0a08-40c4-4d8a-b3e6-4557beb01326 tempest-ImagesOneServerNegativeTestJSON-472538221 tempest-ImagesOneServerNegativeTestJSON-472538221-project-member] Lock "770b6195-ab31-4df0-84b0-e8382732ec32" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 124.613s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 722.673121] env[62460]: DEBUG nova.scheduler.client.report [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 722.879833] env[62460]: INFO nova.compute.manager [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] [instance: 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae] Took 1.05 seconds to deallocate network for instance. [ 723.035466] env[62460]: DEBUG nova.compute.manager [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 723.179743] env[62460]: DEBUG oslo_concurrency.lockutils [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.488s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.180317] env[62460]: DEBUG nova.compute.manager [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 723.183135] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 14.916s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 723.183310] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.183461] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62460) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 723.183761] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.914s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 723.185547] env[62460]: INFO nova.compute.claims [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 723.188842] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff6a4498-f510-4aa4-927a-edff94b3855f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.197479] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77764b79-abfa-433f-b1e9-5eb1c3f61c19 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.214661] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e8050f6-68c0-40de-b6f3-b2ea18693020 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.224432] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1e0be49-57b5-48dc-8939-e882da05e51f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.259746] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181516MB free_disk=135GB free_vcpus=48 pci_devices=None {{(pid=62460) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 723.259942] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 723.556517] env[62460]: DEBUG oslo_concurrency.lockutils [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 723.690034] env[62460]: DEBUG nova.compute.utils [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 723.691782] env[62460]: DEBUG nova.compute.manager [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 723.692009] env[62460]: DEBUG nova.network.neutron [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 723.780125] env[62460]: DEBUG nova.policy [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5f69b3e3f50248ff8e57054294a04686', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0da056d93bdf40c39d6e82e457727ff6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 723.906042] env[62460]: INFO nova.scheduler.client.report [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] Deleted allocations for instance 2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae [ 724.195411] env[62460]: DEBUG nova.compute.manager [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 724.321077] env[62460]: DEBUG nova.network.neutron [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] Successfully created port: 75a13741-f486-4db0-be5b-054880a3cf89 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 724.413103] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d845e65b-94cc-407d-8027-c840c3244176 tempest-ImagesNegativeTestJSON-185559189 tempest-ImagesNegativeTestJSON-185559189-project-member] Lock "2213bb98-48cd-4a9e-8a9b-4b6db30ea2ae" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 114.236s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 724.565681] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e280800-7511-4d8e-af81-263d049ee5e5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.574717] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e8b190a-769f-48a4-a45d-182a32dcde67 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.605701] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57f1e63b-ad8a-4cb7-98fb-667daaf1c515 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.615747] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ba05b76-b2d0-4bf7-a469-f8ee58d217c1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.629926] env[62460]: DEBUG nova.compute.provider_tree [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 724.916484] env[62460]: DEBUG nova.compute.manager [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 725.133816] env[62460]: DEBUG nova.scheduler.client.report [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 725.210574] env[62460]: DEBUG nova.compute.manager [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 725.239362] env[62460]: DEBUG nova.virt.hardware [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 725.239774] env[62460]: DEBUG nova.virt.hardware [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 725.240050] env[62460]: DEBUG nova.virt.hardware [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 725.240291] env[62460]: DEBUG nova.virt.hardware [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 725.240483] env[62460]: DEBUG nova.virt.hardware [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 725.241020] env[62460]: DEBUG nova.virt.hardware [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 725.241020] env[62460]: DEBUG nova.virt.hardware [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 725.241206] env[62460]: DEBUG nova.virt.hardware [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 725.241473] env[62460]: DEBUG nova.virt.hardware [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 725.241689] env[62460]: DEBUG nova.virt.hardware [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 725.241952] env[62460]: DEBUG nova.virt.hardware [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 725.243055] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cd9ff71-7994-4796-9547-2bfc0bc2e926 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.253359] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9238bf41-d5f0-456c-a547-619809c44355 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.440434] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 725.640059] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.456s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 725.640634] env[62460]: DEBUG nova.compute.manager [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 725.643231] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.193s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 725.677267] env[62460]: DEBUG nova.compute.manager [req-205355ac-85b8-41a8-b4ca-54c314a91d1c req-31d9cd27-301b-4015-897f-d35101ea513c service nova] [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] Received event network-changed-75a13741-f486-4db0-be5b-054880a3cf89 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 725.677455] env[62460]: DEBUG nova.compute.manager [req-205355ac-85b8-41a8-b4ca-54c314a91d1c req-31d9cd27-301b-4015-897f-d35101ea513c service nova] [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] Refreshing instance network info cache due to event network-changed-75a13741-f486-4db0-be5b-054880a3cf89. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 725.677669] env[62460]: DEBUG oslo_concurrency.lockutils [req-205355ac-85b8-41a8-b4ca-54c314a91d1c req-31d9cd27-301b-4015-897f-d35101ea513c service nova] Acquiring lock "refresh_cache-ec2297bb-1d29-4ea9-90f7-f4f39716c103" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.677813] env[62460]: DEBUG oslo_concurrency.lockutils [req-205355ac-85b8-41a8-b4ca-54c314a91d1c req-31d9cd27-301b-4015-897f-d35101ea513c service nova] Acquired lock "refresh_cache-ec2297bb-1d29-4ea9-90f7-f4f39716c103" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.677968] env[62460]: DEBUG nova.network.neutron [req-205355ac-85b8-41a8-b4ca-54c314a91d1c req-31d9cd27-301b-4015-897f-d35101ea513c service nova] [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] Refreshing network info cache for port 75a13741-f486-4db0-be5b-054880a3cf89 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 726.149248] env[62460]: DEBUG nova.compute.utils [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 726.153909] env[62460]: DEBUG nova.compute.manager [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 726.158492] env[62460]: DEBUG nova.network.neutron [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 726.239525] env[62460]: DEBUG nova.policy [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '52306a88580742199be8648e4d0da751', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2d44a829041b4112889dbea7ff10d436', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 726.285378] env[62460]: DEBUG nova.network.neutron [req-205355ac-85b8-41a8-b4ca-54c314a91d1c req-31d9cd27-301b-4015-897f-d35101ea513c service nova] [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 726.406144] env[62460]: ERROR nova.compute.manager [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 75a13741-f486-4db0-be5b-054880a3cf89, please check neutron logs for more information. [ 726.406144] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 726.406144] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 726.406144] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 726.406144] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 726.406144] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 726.406144] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 726.406144] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 726.406144] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 726.406144] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 726.406144] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 726.406144] env[62460]: ERROR nova.compute.manager raise self.value [ 726.406144] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 726.406144] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 726.406144] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 726.406144] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 726.407039] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 726.407039] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 726.407039] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 75a13741-f486-4db0-be5b-054880a3cf89, please check neutron logs for more information. [ 726.407039] env[62460]: ERROR nova.compute.manager [ 726.407039] env[62460]: Traceback (most recent call last): [ 726.407039] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 726.407039] env[62460]: listener.cb(fileno) [ 726.407039] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 726.407039] env[62460]: result = function(*args, **kwargs) [ 726.407039] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 726.407039] env[62460]: return func(*args, **kwargs) [ 726.407039] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 726.407039] env[62460]: raise e [ 726.407039] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 726.407039] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 726.407039] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 726.407039] env[62460]: created_port_ids = self._update_ports_for_instance( [ 726.407039] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 726.407039] env[62460]: with excutils.save_and_reraise_exception(): [ 726.407039] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 726.407039] env[62460]: self.force_reraise() [ 726.407039] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 726.407039] env[62460]: raise self.value [ 726.407039] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 726.407039] env[62460]: updated_port = self._update_port( [ 726.407039] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 726.407039] env[62460]: _ensure_no_port_binding_failure(port) [ 726.407039] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 726.407039] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 726.408605] env[62460]: nova.exception.PortBindingFailed: Binding failed for port 75a13741-f486-4db0-be5b-054880a3cf89, please check neutron logs for more information. [ 726.408605] env[62460]: Removing descriptor: 19 [ 726.408605] env[62460]: ERROR nova.compute.manager [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 75a13741-f486-4db0-be5b-054880a3cf89, please check neutron logs for more information. [ 726.408605] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] Traceback (most recent call last): [ 726.408605] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 726.408605] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] yield resources [ 726.408605] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 726.408605] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] self.driver.spawn(context, instance, image_meta, [ 726.408605] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 726.408605] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] self._vmops.spawn(context, instance, image_meta, injected_files, [ 726.408605] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 726.408605] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] vm_ref = self.build_virtual_machine(instance, [ 726.408998] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 726.408998] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] vif_infos = vmwarevif.get_vif_info(self._session, [ 726.408998] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 726.408998] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] for vif in network_info: [ 726.408998] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 726.408998] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] return self._sync_wrapper(fn, *args, **kwargs) [ 726.408998] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 726.408998] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] self.wait() [ 726.408998] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 726.408998] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] self[:] = self._gt.wait() [ 726.408998] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 726.408998] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] return self._exit_event.wait() [ 726.408998] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 726.409370] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] result = hub.switch() [ 726.409370] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 726.409370] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] return self.greenlet.switch() [ 726.409370] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 726.409370] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] result = function(*args, **kwargs) [ 726.409370] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 726.409370] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] return func(*args, **kwargs) [ 726.409370] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 726.409370] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] raise e [ 726.409370] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 726.409370] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] nwinfo = self.network_api.allocate_for_instance( [ 726.409370] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 726.409370] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] created_port_ids = self._update_ports_for_instance( [ 726.409759] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 726.409759] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] with excutils.save_and_reraise_exception(): [ 726.409759] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 726.409759] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] self.force_reraise() [ 726.409759] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 726.409759] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] raise self.value [ 726.409759] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 726.409759] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] updated_port = self._update_port( [ 726.409759] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 726.409759] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] _ensure_no_port_binding_failure(port) [ 726.409759] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 726.409759] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] raise exception.PortBindingFailed(port_id=port['id']) [ 726.411615] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] nova.exception.PortBindingFailed: Binding failed for port 75a13741-f486-4db0-be5b-054880a3cf89, please check neutron logs for more information. [ 726.411615] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] [ 726.411615] env[62460]: INFO nova.compute.manager [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] Terminating instance [ 726.411841] env[62460]: DEBUG oslo_concurrency.lockutils [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "refresh_cache-ec2297bb-1d29-4ea9-90f7-f4f39716c103" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.515338] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-977da2c9-be24-41d7-9220-0adc1c625845 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.525036] env[62460]: DEBUG nova.network.neutron [req-205355ac-85b8-41a8-b4ca-54c314a91d1c req-31d9cd27-301b-4015-897f-d35101ea513c service nova] [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.527699] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3d66416-0464-4675-a28d-0a1221871448 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.569223] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a837ce67-1a2b-49d7-97e0-fea1e3fa411f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.576307] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d502ad0d-d1a6-4d07-b11a-b7abc04f1ad2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.591852] env[62460]: DEBUG nova.compute.provider_tree [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 726.663276] env[62460]: DEBUG nova.compute.manager [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 726.788949] env[62460]: DEBUG nova.network.neutron [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] Successfully created port: ed57313b-d260-4e7a-99b1-e4a6318a31b5 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 727.032012] env[62460]: DEBUG oslo_concurrency.lockutils [req-205355ac-85b8-41a8-b4ca-54c314a91d1c req-31d9cd27-301b-4015-897f-d35101ea513c service nova] Releasing lock "refresh_cache-ec2297bb-1d29-4ea9-90f7-f4f39716c103" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.032561] env[62460]: DEBUG oslo_concurrency.lockutils [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquired lock "refresh_cache-ec2297bb-1d29-4ea9-90f7-f4f39716c103" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.032837] env[62460]: DEBUG nova.network.neutron [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 727.095447] env[62460]: DEBUG nova.scheduler.client.report [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 727.564202] env[62460]: DEBUG nova.network.neutron [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 727.603576] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.960s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 727.604456] env[62460]: ERROR nova.compute.manager [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2d5fe2cf-85d6-4ff8-aa4e-a8f56a2c2971, please check neutron logs for more information. [ 727.604456] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] Traceback (most recent call last): [ 727.604456] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 727.604456] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] self.driver.spawn(context, instance, image_meta, [ 727.604456] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 727.604456] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] self._vmops.spawn(context, instance, image_meta, injected_files, [ 727.604456] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 727.604456] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] vm_ref = self.build_virtual_machine(instance, [ 727.604456] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 727.604456] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] vif_infos = vmwarevif.get_vif_info(self._session, [ 727.604456] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 727.604923] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] for vif in network_info: [ 727.604923] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 727.604923] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] return self._sync_wrapper(fn, *args, **kwargs) [ 727.604923] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 727.604923] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] self.wait() [ 727.604923] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 727.604923] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] self[:] = self._gt.wait() [ 727.604923] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 727.604923] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] return self._exit_event.wait() [ 727.604923] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 727.604923] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] result = hub.switch() [ 727.604923] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 727.604923] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] return self.greenlet.switch() [ 727.605336] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 727.605336] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] result = function(*args, **kwargs) [ 727.605336] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 727.605336] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] return func(*args, **kwargs) [ 727.605336] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 727.605336] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] raise e [ 727.605336] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 727.605336] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] nwinfo = self.network_api.allocate_for_instance( [ 727.605336] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 727.605336] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] created_port_ids = self._update_ports_for_instance( [ 727.605336] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 727.605336] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] with excutils.save_and_reraise_exception(): [ 727.605336] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 727.605879] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] self.force_reraise() [ 727.605879] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 727.605879] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] raise self.value [ 727.605879] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 727.605879] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] updated_port = self._update_port( [ 727.605879] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 727.605879] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] _ensure_no_port_binding_failure(port) [ 727.605879] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 727.605879] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] raise exception.PortBindingFailed(port_id=port['id']) [ 727.605879] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] nova.exception.PortBindingFailed: Binding failed for port 2d5fe2cf-85d6-4ff8-aa4e-a8f56a2c2971, please check neutron logs for more information. [ 727.605879] env[62460]: ERROR nova.compute.manager [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] [ 727.606252] env[62460]: DEBUG nova.compute.utils [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] Binding failed for port 2d5fe2cf-85d6-4ff8-aa4e-a8f56a2c2971, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 727.608088] env[62460]: DEBUG nova.compute.manager [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] Build of instance c37f1b55-7a06-4c93-9492-d2bd6d2bfe83 was re-scheduled: Binding failed for port 2d5fe2cf-85d6-4ff8-aa4e-a8f56a2c2971, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 727.608518] env[62460]: DEBUG nova.compute.manager [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 727.608760] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] Acquiring lock "refresh_cache-c37f1b55-7a06-4c93-9492-d2bd6d2bfe83" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 727.608939] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] Acquired lock "refresh_cache-c37f1b55-7a06-4c93-9492-d2bd6d2bfe83" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.609134] env[62460]: DEBUG nova.network.neutron [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 727.610283] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.824s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 727.611931] env[62460]: INFO nova.compute.claims [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 727.682386] env[62460]: DEBUG nova.compute.manager [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 727.727751] env[62460]: DEBUG nova.virt.hardware [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 727.728048] env[62460]: DEBUG nova.virt.hardware [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 727.728220] env[62460]: DEBUG nova.virt.hardware [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 727.728405] env[62460]: DEBUG nova.virt.hardware [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 727.728553] env[62460]: DEBUG nova.virt.hardware [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 727.728701] env[62460]: DEBUG nova.virt.hardware [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 727.728911] env[62460]: DEBUG nova.virt.hardware [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 727.730735] env[62460]: DEBUG nova.virt.hardware [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 727.731515] env[62460]: DEBUG nova.virt.hardware [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 727.731739] env[62460]: DEBUG nova.virt.hardware [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 727.732026] env[62460]: DEBUG nova.virt.hardware [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 727.732936] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15f3ab4a-fa0e-45be-97a2-ab75b3215d15 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.744893] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-354b3715-e660-4838-ad4c-af1c568feece {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.750367] env[62460]: DEBUG nova.compute.manager [req-e63913fa-a5a8-4a78-b1ba-a1fcce1a614f req-7a7ce4c9-e6d1-4cca-ad07-1bd52d80b91c service nova] [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] Received event network-vif-deleted-75a13741-f486-4db0-be5b-054880a3cf89 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 727.776363] env[62460]: DEBUG nova.network.neutron [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.986316] env[62460]: DEBUG nova.compute.manager [req-d84e1517-73ed-423d-8f52-b15352a8cd4e req-1b954968-ec23-4c4f-9635-45d7599158df service nova] [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] Received event network-changed-ed57313b-d260-4e7a-99b1-e4a6318a31b5 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 727.986498] env[62460]: DEBUG nova.compute.manager [req-d84e1517-73ed-423d-8f52-b15352a8cd4e req-1b954968-ec23-4c4f-9635-45d7599158df service nova] [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] Refreshing instance network info cache due to event network-changed-ed57313b-d260-4e7a-99b1-e4a6318a31b5. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 727.986707] env[62460]: DEBUG oslo_concurrency.lockutils [req-d84e1517-73ed-423d-8f52-b15352a8cd4e req-1b954968-ec23-4c4f-9635-45d7599158df service nova] Acquiring lock "refresh_cache-0e00ca48-f9e1-4999-aad5-d7965f7ddc28" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 727.987062] env[62460]: DEBUG oslo_concurrency.lockutils [req-d84e1517-73ed-423d-8f52-b15352a8cd4e req-1b954968-ec23-4c4f-9635-45d7599158df service nova] Acquired lock "refresh_cache-0e00ca48-f9e1-4999-aad5-d7965f7ddc28" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.987062] env[62460]: DEBUG nova.network.neutron [req-d84e1517-73ed-423d-8f52-b15352a8cd4e req-1b954968-ec23-4c4f-9635-45d7599158df service nova] [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] Refreshing network info cache for port ed57313b-d260-4e7a-99b1-e4a6318a31b5 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 728.144411] env[62460]: DEBUG nova.network.neutron [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 728.231424] env[62460]: DEBUG nova.network.neutron [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.270481] env[62460]: ERROR nova.compute.manager [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ed57313b-d260-4e7a-99b1-e4a6318a31b5, please check neutron logs for more information. [ 728.270481] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 728.270481] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 728.270481] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 728.270481] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 728.270481] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 728.270481] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 728.270481] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 728.270481] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 728.270481] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 728.270481] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 728.270481] env[62460]: ERROR nova.compute.manager raise self.value [ 728.270481] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 728.270481] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 728.270481] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 728.270481] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 728.271013] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 728.271013] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 728.271013] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ed57313b-d260-4e7a-99b1-e4a6318a31b5, please check neutron logs for more information. [ 728.271013] env[62460]: ERROR nova.compute.manager [ 728.271013] env[62460]: Traceback (most recent call last): [ 728.271013] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 728.271013] env[62460]: listener.cb(fileno) [ 728.271013] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 728.271013] env[62460]: result = function(*args, **kwargs) [ 728.271013] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 728.271013] env[62460]: return func(*args, **kwargs) [ 728.271013] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 728.271013] env[62460]: raise e [ 728.271013] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 728.271013] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 728.271013] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 728.271013] env[62460]: created_port_ids = self._update_ports_for_instance( [ 728.271013] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 728.271013] env[62460]: with excutils.save_and_reraise_exception(): [ 728.271013] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 728.271013] env[62460]: self.force_reraise() [ 728.271013] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 728.271013] env[62460]: raise self.value [ 728.271013] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 728.271013] env[62460]: updated_port = self._update_port( [ 728.271013] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 728.271013] env[62460]: _ensure_no_port_binding_failure(port) [ 728.271013] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 728.271013] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 728.271871] env[62460]: nova.exception.PortBindingFailed: Binding failed for port ed57313b-d260-4e7a-99b1-e4a6318a31b5, please check neutron logs for more information. [ 728.271871] env[62460]: Removing descriptor: 18 [ 728.271945] env[62460]: ERROR nova.compute.manager [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ed57313b-d260-4e7a-99b1-e4a6318a31b5, please check neutron logs for more information. [ 728.271945] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] Traceback (most recent call last): [ 728.271945] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 728.271945] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] yield resources [ 728.271945] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 728.271945] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] self.driver.spawn(context, instance, image_meta, [ 728.271945] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 728.271945] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] self._vmops.spawn(context, instance, image_meta, injected_files, [ 728.271945] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 728.271945] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] vm_ref = self.build_virtual_machine(instance, [ 728.271945] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 728.272310] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] vif_infos = vmwarevif.get_vif_info(self._session, [ 728.272310] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 728.272310] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] for vif in network_info: [ 728.272310] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 728.272310] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] return self._sync_wrapper(fn, *args, **kwargs) [ 728.272310] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 728.272310] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] self.wait() [ 728.272310] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 728.272310] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] self[:] = self._gt.wait() [ 728.272310] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 728.272310] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] return self._exit_event.wait() [ 728.272310] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 728.272310] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] result = hub.switch() [ 728.272815] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 728.272815] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] return self.greenlet.switch() [ 728.272815] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 728.272815] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] result = function(*args, **kwargs) [ 728.272815] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 728.272815] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] return func(*args, **kwargs) [ 728.272815] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 728.272815] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] raise e [ 728.272815] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 728.272815] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] nwinfo = self.network_api.allocate_for_instance( [ 728.272815] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 728.272815] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] created_port_ids = self._update_ports_for_instance( [ 728.272815] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 728.273224] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] with excutils.save_and_reraise_exception(): [ 728.273224] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 728.273224] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] self.force_reraise() [ 728.273224] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 728.273224] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] raise self.value [ 728.273224] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 728.273224] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] updated_port = self._update_port( [ 728.273224] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 728.273224] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] _ensure_no_port_binding_failure(port) [ 728.273224] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 728.273224] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] raise exception.PortBindingFailed(port_id=port['id']) [ 728.273224] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] nova.exception.PortBindingFailed: Binding failed for port ed57313b-d260-4e7a-99b1-e4a6318a31b5, please check neutron logs for more information. [ 728.273224] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] [ 728.273655] env[62460]: INFO nova.compute.manager [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] Terminating instance [ 728.276949] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] Acquiring lock "refresh_cache-0e00ca48-f9e1-4999-aad5-d7965f7ddc28" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.278966] env[62460]: DEBUG oslo_concurrency.lockutils [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Releasing lock "refresh_cache-ec2297bb-1d29-4ea9-90f7-f4f39716c103" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 728.279604] env[62460]: DEBUG nova.compute.manager [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 728.279809] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 728.280368] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8b122055-0be1-4f9f-a666-3e60bbfd5273 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.290934] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-214bf781-1dbf-4140-bc56-3874bfda07a9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.315984] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ec2297bb-1d29-4ea9-90f7-f4f39716c103 could not be found. [ 728.316421] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 728.316514] env[62460]: INFO nova.compute.manager [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] Took 0.04 seconds to destroy the instance on the hypervisor. [ 728.316786] env[62460]: DEBUG oslo.service.loopingcall [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 728.317115] env[62460]: DEBUG nova.compute.manager [-] [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 728.317220] env[62460]: DEBUG nova.network.neutron [-] [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 728.334858] env[62460]: DEBUG nova.network.neutron [-] [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 728.521514] env[62460]: DEBUG nova.network.neutron [req-d84e1517-73ed-423d-8f52-b15352a8cd4e req-1b954968-ec23-4c4f-9635-45d7599158df service nova] [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 728.613958] env[62460]: DEBUG nova.network.neutron [req-d84e1517-73ed-423d-8f52-b15352a8cd4e req-1b954968-ec23-4c4f-9635-45d7599158df service nova] [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.735117] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] Releasing lock "refresh_cache-c37f1b55-7a06-4c93-9492-d2bd6d2bfe83" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 728.735436] env[62460]: DEBUG nova.compute.manager [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 728.735625] env[62460]: DEBUG nova.compute.manager [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 728.735804] env[62460]: DEBUG nova.network.neutron [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 728.762807] env[62460]: DEBUG nova.network.neutron [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 728.838384] env[62460]: DEBUG nova.network.neutron [-] [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.997261] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88ff28f8-8cdf-4cf0-8307-5bd081f92585 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.007476] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efab2867-5003-49d7-8bca-5f1bbea3d7a8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.041543] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e175c842-b321-47df-8aa0-5ca0b67d5b01 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.050593] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-696a83a9-77e6-4184-aa5e-0f4ad2c26798 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.066131] env[62460]: DEBUG nova.compute.provider_tree [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 729.116255] env[62460]: DEBUG oslo_concurrency.lockutils [req-d84e1517-73ed-423d-8f52-b15352a8cd4e req-1b954968-ec23-4c4f-9635-45d7599158df service nova] Releasing lock "refresh_cache-0e00ca48-f9e1-4999-aad5-d7965f7ddc28" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 729.116677] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] Acquired lock "refresh_cache-0e00ca48-f9e1-4999-aad5-d7965f7ddc28" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.116899] env[62460]: DEBUG nova.network.neutron [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 729.265435] env[62460]: DEBUG nova.network.neutron [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.345787] env[62460]: INFO nova.compute.manager [-] [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] Took 1.03 seconds to deallocate network for instance. [ 729.347349] env[62460]: DEBUG nova.compute.claims [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 729.347349] env[62460]: DEBUG oslo_concurrency.lockutils [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 729.412069] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "b76f37a0-91d0-4a01-9d95-9c6586081175" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 729.412201] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "b76f37a0-91d0-4a01-9d95-9c6586081175" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 729.573306] env[62460]: DEBUG nova.scheduler.client.report [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 729.653935] env[62460]: DEBUG nova.network.neutron [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 729.769625] env[62460]: INFO nova.compute.manager [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] [instance: c37f1b55-7a06-4c93-9492-d2bd6d2bfe83] Took 1.03 seconds to deallocate network for instance. [ 729.779629] env[62460]: DEBUG nova.network.neutron [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 730.026620] env[62460]: DEBUG nova.compute.manager [req-bd790060-ebb0-4f19-9ec0-91a2f5f45838 req-f71f6a71-cebd-4422-b6c3-f688197ab439 service nova] [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] Received event network-vif-deleted-ed57313b-d260-4e7a-99b1-e4a6318a31b5 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 730.080506] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.470s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 730.081456] env[62460]: DEBUG nova.compute.manager [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 730.087127] env[62460]: DEBUG oslo_concurrency.lockutils [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.419s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 730.091116] env[62460]: INFO nova.compute.claims [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 730.286133] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] Releasing lock "refresh_cache-0e00ca48-f9e1-4999-aad5-d7965f7ddc28" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 730.286133] env[62460]: DEBUG nova.compute.manager [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 730.286329] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 730.287060] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-927d3464-a331-4a48-8a98-d7e09a8401a3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.296050] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82a186d5-8a30-44c9-a749-427c5f393767 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.325377] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0e00ca48-f9e1-4999-aad5-d7965f7ddc28 could not be found. [ 730.325734] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 730.325849] env[62460]: INFO nova.compute.manager [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] Took 0.04 seconds to destroy the instance on the hypervisor. [ 730.326161] env[62460]: DEBUG oslo.service.loopingcall [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 730.326652] env[62460]: DEBUG nova.compute.manager [-] [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 730.326782] env[62460]: DEBUG nova.network.neutron [-] [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 730.347103] env[62460]: DEBUG nova.network.neutron [-] [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 730.594282] env[62460]: DEBUG nova.compute.utils [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 730.597823] env[62460]: DEBUG nova.compute.manager [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 730.601230] env[62460]: DEBUG nova.network.neutron [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 730.671020] env[62460]: DEBUG nova.policy [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '93019ed5bc5848d68c3c225032a01674', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'de568d90eb94456aa0714fa952b53b75', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 730.803400] env[62460]: INFO nova.scheduler.client.report [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] Deleted allocations for instance c37f1b55-7a06-4c93-9492-d2bd6d2bfe83 [ 730.853569] env[62460]: DEBUG nova.network.neutron [-] [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.019793] env[62460]: DEBUG nova.network.neutron [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] Successfully created port: 319af934-2bcb-4020-bc62-7039d1fc7eb4 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 731.099774] env[62460]: DEBUG nova.compute.manager [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 731.252708] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Acquiring lock "5b6e8205-003e-49c0-a73d-be2e032a8272" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 731.252708] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lock "5b6e8205-003e-49c0-a73d-be2e032a8272" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 731.316082] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4d39d144-870e-45a1-8241-19aa3a8108af tempest-TenantUsagesTestJSON-1511548734 tempest-TenantUsagesTestJSON-1511548734-project-member] Lock "c37f1b55-7a06-4c93-9492-d2bd6d2bfe83" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 120.381s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 731.356237] env[62460]: INFO nova.compute.manager [-] [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] Took 1.03 seconds to deallocate network for instance. [ 731.360719] env[62460]: DEBUG nova.compute.claims [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 731.360958] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 731.469929] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a3ad0de-5f79-435e-aaaf-1286c32abce3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.479509] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0933109a-cb93-4bc4-9199-a053f39c9e6d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.513303] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5bf74f9-b682-4ab3-8bc7-5785d3502897 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.521271] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a3ff0dd-e4dd-48e7-93b7-9ea59c9d0f06 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.535399] env[62460]: DEBUG nova.compute.provider_tree [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 731.819877] env[62460]: DEBUG nova.compute.manager [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 732.039547] env[62460]: DEBUG nova.scheduler.client.report [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 732.109030] env[62460]: DEBUG nova.compute.manager [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 732.155692] env[62460]: DEBUG nova.virt.hardware [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 732.157677] env[62460]: DEBUG nova.virt.hardware [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 732.157677] env[62460]: DEBUG nova.virt.hardware [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 732.157677] env[62460]: DEBUG nova.virt.hardware [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 732.157677] env[62460]: DEBUG nova.virt.hardware [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 732.157677] env[62460]: DEBUG nova.virt.hardware [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 732.157904] env[62460]: DEBUG nova.virt.hardware [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 732.157904] env[62460]: DEBUG nova.virt.hardware [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 732.157904] env[62460]: DEBUG nova.virt.hardware [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 732.157904] env[62460]: DEBUG nova.virt.hardware [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 732.157904] env[62460]: DEBUG nova.virt.hardware [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 732.161552] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d7ce943-c75d-42c1-aaf7-04f799e9d00b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.171544] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4634a3ad-0357-4877-9b60-f1193a045df3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.224829] env[62460]: DEBUG nova.compute.manager [req-88a8e7b6-0987-412f-af70-82177fdce015 req-05cb1b47-449e-4d2e-94b4-1542493b64c4 service nova] [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] Received event network-changed-319af934-2bcb-4020-bc62-7039d1fc7eb4 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 732.224829] env[62460]: DEBUG nova.compute.manager [req-88a8e7b6-0987-412f-af70-82177fdce015 req-05cb1b47-449e-4d2e-94b4-1542493b64c4 service nova] [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] Refreshing instance network info cache due to event network-changed-319af934-2bcb-4020-bc62-7039d1fc7eb4. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 732.224829] env[62460]: DEBUG oslo_concurrency.lockutils [req-88a8e7b6-0987-412f-af70-82177fdce015 req-05cb1b47-449e-4d2e-94b4-1542493b64c4 service nova] Acquiring lock "refresh_cache-ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.225103] env[62460]: DEBUG oslo_concurrency.lockutils [req-88a8e7b6-0987-412f-af70-82177fdce015 req-05cb1b47-449e-4d2e-94b4-1542493b64c4 service nova] Acquired lock "refresh_cache-ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.226913] env[62460]: DEBUG nova.network.neutron [req-88a8e7b6-0987-412f-af70-82177fdce015 req-05cb1b47-449e-4d2e-94b4-1542493b64c4 service nova] [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] Refreshing network info cache for port 319af934-2bcb-4020-bc62-7039d1fc7eb4 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 732.294487] env[62460]: ERROR nova.compute.manager [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 319af934-2bcb-4020-bc62-7039d1fc7eb4, please check neutron logs for more information. [ 732.294487] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 732.294487] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 732.294487] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 732.294487] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 732.294487] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 732.294487] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 732.294487] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 732.294487] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 732.294487] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 732.294487] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 732.294487] env[62460]: ERROR nova.compute.manager raise self.value [ 732.294487] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 732.294487] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 732.294487] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 732.294487] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 732.294973] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 732.294973] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 732.294973] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 319af934-2bcb-4020-bc62-7039d1fc7eb4, please check neutron logs for more information. [ 732.294973] env[62460]: ERROR nova.compute.manager [ 732.294973] env[62460]: Traceback (most recent call last): [ 732.294973] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 732.294973] env[62460]: listener.cb(fileno) [ 732.294973] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 732.294973] env[62460]: result = function(*args, **kwargs) [ 732.294973] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 732.294973] env[62460]: return func(*args, **kwargs) [ 732.294973] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 732.294973] env[62460]: raise e [ 732.294973] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 732.294973] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 732.294973] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 732.294973] env[62460]: created_port_ids = self._update_ports_for_instance( [ 732.294973] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 732.294973] env[62460]: with excutils.save_and_reraise_exception(): [ 732.294973] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 732.294973] env[62460]: self.force_reraise() [ 732.294973] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 732.294973] env[62460]: raise self.value [ 732.294973] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 732.294973] env[62460]: updated_port = self._update_port( [ 732.294973] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 732.294973] env[62460]: _ensure_no_port_binding_failure(port) [ 732.294973] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 732.294973] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 732.295973] env[62460]: nova.exception.PortBindingFailed: Binding failed for port 319af934-2bcb-4020-bc62-7039d1fc7eb4, please check neutron logs for more information. [ 732.295973] env[62460]: Removing descriptor: 18 [ 732.295973] env[62460]: ERROR nova.compute.manager [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 319af934-2bcb-4020-bc62-7039d1fc7eb4, please check neutron logs for more information. [ 732.295973] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] Traceback (most recent call last): [ 732.295973] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 732.295973] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] yield resources [ 732.295973] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 732.295973] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] self.driver.spawn(context, instance, image_meta, [ 732.295973] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 732.295973] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] self._vmops.spawn(context, instance, image_meta, injected_files, [ 732.295973] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 732.295973] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] vm_ref = self.build_virtual_machine(instance, [ 732.296349] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 732.296349] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] vif_infos = vmwarevif.get_vif_info(self._session, [ 732.296349] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 732.296349] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] for vif in network_info: [ 732.296349] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 732.296349] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] return self._sync_wrapper(fn, *args, **kwargs) [ 732.296349] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 732.296349] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] self.wait() [ 732.296349] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 732.296349] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] self[:] = self._gt.wait() [ 732.296349] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 732.296349] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] return self._exit_event.wait() [ 732.296349] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 732.297169] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] result = hub.switch() [ 732.297169] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 732.297169] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] return self.greenlet.switch() [ 732.297169] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 732.297169] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] result = function(*args, **kwargs) [ 732.297169] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 732.297169] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] return func(*args, **kwargs) [ 732.297169] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 732.297169] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] raise e [ 732.297169] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 732.297169] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] nwinfo = self.network_api.allocate_for_instance( [ 732.297169] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 732.297169] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] created_port_ids = self._update_ports_for_instance( [ 732.297573] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 732.297573] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] with excutils.save_and_reraise_exception(): [ 732.297573] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 732.297573] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] self.force_reraise() [ 732.297573] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 732.297573] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] raise self.value [ 732.297573] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 732.297573] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] updated_port = self._update_port( [ 732.297573] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 732.297573] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] _ensure_no_port_binding_failure(port) [ 732.297573] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 732.297573] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] raise exception.PortBindingFailed(port_id=port['id']) [ 732.297942] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] nova.exception.PortBindingFailed: Binding failed for port 319af934-2bcb-4020-bc62-7039d1fc7eb4, please check neutron logs for more information. [ 732.297942] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] [ 732.297942] env[62460]: INFO nova.compute.manager [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] Terminating instance [ 732.298302] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] Acquiring lock "refresh_cache-ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.341574] env[62460]: DEBUG oslo_concurrency.lockutils [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 732.544504] env[62460]: DEBUG oslo_concurrency.lockutils [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.457s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 732.545137] env[62460]: DEBUG nova.compute.manager [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 732.547849] env[62460]: DEBUG oslo_concurrency.lockutils [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.915s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 732.549393] env[62460]: INFO nova.compute.claims [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 4ed90d16-81a6-4dbd-8936-0e137151171f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 732.758142] env[62460]: DEBUG nova.network.neutron [req-88a8e7b6-0987-412f-af70-82177fdce015 req-05cb1b47-449e-4d2e-94b4-1542493b64c4 service nova] [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 732.991036] env[62460]: DEBUG nova.network.neutron [req-88a8e7b6-0987-412f-af70-82177fdce015 req-05cb1b47-449e-4d2e-94b4-1542493b64c4 service nova] [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.056499] env[62460]: DEBUG nova.compute.utils [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 733.059587] env[62460]: DEBUG nova.compute.manager [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 733.063257] env[62460]: DEBUG nova.network.neutron [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 733.131604] env[62460]: DEBUG nova.policy [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0e6f85699c4344609bc13644d34609cc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cf6954e6a10d4219b35ea3e62f95ace0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 733.488220] env[62460]: DEBUG nova.network.neutron [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] Successfully created port: 6fc7ca35-5bb9-4d46-9356-ceb957531a76 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 733.493370] env[62460]: DEBUG oslo_concurrency.lockutils [req-88a8e7b6-0987-412f-af70-82177fdce015 req-05cb1b47-449e-4d2e-94b4-1542493b64c4 service nova] Releasing lock "refresh_cache-ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.493832] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] Acquired lock "refresh_cache-ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 733.494130] env[62460]: DEBUG nova.network.neutron [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 733.560100] env[62460]: DEBUG nova.compute.manager [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 733.936773] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Acquiring lock "9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.937063] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Lock "9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.964799] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Acquiring lock "0269dc64-d2b1-43c5-bdf7-11d97e534819" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.964799] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Lock "0269dc64-d2b1-43c5-bdf7-11d97e534819" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.968910] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53b1df3a-038b-41e0-af79-b3daaf7bcd29 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.977560] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4379662c-54cf-4ee5-840b-b05c9e70f032 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.011344] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19441d93-fec7-40d8-a29c-960a3986be6c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.021022] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c24c3206-a975-42f3-87e1-fa1c5e84e82c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.037803] env[62460]: DEBUG nova.compute.provider_tree [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 734.041182] env[62460]: DEBUG nova.network.neutron [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 734.137130] env[62460]: DEBUG nova.network.neutron [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.255298] env[62460]: DEBUG nova.compute.manager [req-c37c5b31-a992-4d50-9c76-b375c72ce336 req-103f4da9-288b-429f-a0a3-b364bdc2b060 service nova] [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] Received event network-vif-deleted-319af934-2bcb-4020-bc62-7039d1fc7eb4 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 734.543105] env[62460]: DEBUG nova.scheduler.client.report [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 734.570981] env[62460]: DEBUG nova.compute.manager [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 734.578061] env[62460]: ERROR nova.compute.manager [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6fc7ca35-5bb9-4d46-9356-ceb957531a76, please check neutron logs for more information. [ 734.578061] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 734.578061] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 734.578061] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 734.578061] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 734.578061] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 734.578061] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 734.578061] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 734.578061] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 734.578061] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 734.578061] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 734.578061] env[62460]: ERROR nova.compute.manager raise self.value [ 734.578061] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 734.578061] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 734.578061] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 734.578061] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 734.578637] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 734.578637] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 734.578637] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6fc7ca35-5bb9-4d46-9356-ceb957531a76, please check neutron logs for more information. [ 734.578637] env[62460]: ERROR nova.compute.manager [ 734.578637] env[62460]: Traceback (most recent call last): [ 734.578637] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 734.578637] env[62460]: listener.cb(fileno) [ 734.578637] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 734.578637] env[62460]: result = function(*args, **kwargs) [ 734.578637] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 734.578637] env[62460]: return func(*args, **kwargs) [ 734.578637] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 734.578637] env[62460]: raise e [ 734.578637] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 734.578637] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 734.578637] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 734.578637] env[62460]: created_port_ids = self._update_ports_for_instance( [ 734.578637] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 734.578637] env[62460]: with excutils.save_and_reraise_exception(): [ 734.578637] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 734.578637] env[62460]: self.force_reraise() [ 734.578637] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 734.578637] env[62460]: raise self.value [ 734.578637] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 734.578637] env[62460]: updated_port = self._update_port( [ 734.578637] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 734.578637] env[62460]: _ensure_no_port_binding_failure(port) [ 734.578637] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 734.578637] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 734.579518] env[62460]: nova.exception.PortBindingFailed: Binding failed for port 6fc7ca35-5bb9-4d46-9356-ceb957531a76, please check neutron logs for more information. [ 734.579518] env[62460]: Removing descriptor: 18 [ 734.579713] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Acquiring lock "9d1a5830-f3c0-4d18-9338-16f7b6962c6a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 734.579965] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Lock "9d1a5830-f3c0-4d18-9338-16f7b6962c6a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 734.599199] env[62460]: DEBUG nova.virt.hardware [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 734.599199] env[62460]: DEBUG nova.virt.hardware [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 734.599199] env[62460]: DEBUG nova.virt.hardware [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 734.599352] env[62460]: DEBUG nova.virt.hardware [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 734.599352] env[62460]: DEBUG nova.virt.hardware [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 734.599352] env[62460]: DEBUG nova.virt.hardware [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 734.599352] env[62460]: DEBUG nova.virt.hardware [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 734.599352] env[62460]: DEBUG nova.virt.hardware [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 734.599508] env[62460]: DEBUG nova.virt.hardware [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 734.599508] env[62460]: DEBUG nova.virt.hardware [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 734.599508] env[62460]: DEBUG nova.virt.hardware [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 734.601415] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fbe9d92-94d3-4088-bdc0-c71fb20d970a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.609359] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49858d09-8caf-4aa8-b405-7811e4d8edb0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.623351] env[62460]: ERROR nova.compute.manager [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6fc7ca35-5bb9-4d46-9356-ceb957531a76, please check neutron logs for more information. [ 734.623351] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] Traceback (most recent call last): [ 734.623351] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 734.623351] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] yield resources [ 734.623351] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 734.623351] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] self.driver.spawn(context, instance, image_meta, [ 734.623351] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 734.623351] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 734.623351] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 734.623351] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] vm_ref = self.build_virtual_machine(instance, [ 734.623351] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 734.623793] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] vif_infos = vmwarevif.get_vif_info(self._session, [ 734.623793] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 734.623793] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] for vif in network_info: [ 734.623793] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 734.623793] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] return self._sync_wrapper(fn, *args, **kwargs) [ 734.623793] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 734.623793] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] self.wait() [ 734.623793] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 734.623793] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] self[:] = self._gt.wait() [ 734.623793] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 734.623793] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] return self._exit_event.wait() [ 734.623793] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 734.623793] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] current.throw(*self._exc) [ 734.624244] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 734.624244] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] result = function(*args, **kwargs) [ 734.624244] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 734.624244] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] return func(*args, **kwargs) [ 734.624244] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 734.624244] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] raise e [ 734.624244] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 734.624244] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] nwinfo = self.network_api.allocate_for_instance( [ 734.624244] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 734.624244] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] created_port_ids = self._update_ports_for_instance( [ 734.624244] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 734.624244] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] with excutils.save_and_reraise_exception(): [ 734.624244] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 734.624677] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] self.force_reraise() [ 734.624677] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 734.624677] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] raise self.value [ 734.624677] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 734.624677] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] updated_port = self._update_port( [ 734.624677] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 734.624677] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] _ensure_no_port_binding_failure(port) [ 734.624677] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 734.624677] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] raise exception.PortBindingFailed(port_id=port['id']) [ 734.624677] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] nova.exception.PortBindingFailed: Binding failed for port 6fc7ca35-5bb9-4d46-9356-ceb957531a76, please check neutron logs for more information. [ 734.624677] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] [ 734.624677] env[62460]: INFO nova.compute.manager [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] Terminating instance [ 734.626102] env[62460]: DEBUG oslo_concurrency.lockutils [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] Acquiring lock "refresh_cache-c4c5abc4-a5ed-4337-a930-d37d579819f8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 734.626102] env[62460]: DEBUG oslo_concurrency.lockutils [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] Acquired lock "refresh_cache-c4c5abc4-a5ed-4337-a930-d37d579819f8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.626224] env[62460]: DEBUG nova.network.neutron [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 734.639260] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] Releasing lock "refresh_cache-ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 734.639463] env[62460]: DEBUG nova.compute.manager [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 734.639659] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 734.639931] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dd99290c-b274-42c9-aba6-3cd49aa847c1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.649124] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3ee7b02-f87a-48a1-96d2-3ea489ee3bd9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.672102] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee could not be found. [ 734.672472] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 734.672778] env[62460]: INFO nova.compute.manager [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] Took 0.03 seconds to destroy the instance on the hypervisor. [ 734.673212] env[62460]: DEBUG oslo.service.loopingcall [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 734.673452] env[62460]: DEBUG nova.compute.manager [-] [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 734.673556] env[62460]: DEBUG nova.network.neutron [-] [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 734.687832] env[62460]: DEBUG nova.network.neutron [-] [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 735.050229] env[62460]: DEBUG oslo_concurrency.lockutils [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.502s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 735.050868] env[62460]: DEBUG nova.compute.manager [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 4ed90d16-81a6-4dbd-8936-0e137151171f] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 735.053459] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.450s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 735.054852] env[62460]: INFO nova.compute.claims [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 735.144182] env[62460]: DEBUG nova.network.neutron [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 735.191268] env[62460]: DEBUG nova.network.neutron [-] [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.230583] env[62460]: DEBUG nova.network.neutron [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.559016] env[62460]: DEBUG nova.compute.utils [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 735.563056] env[62460]: DEBUG nova.compute.manager [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 4ed90d16-81a6-4dbd-8936-0e137151171f] Not allocating networking since 'none' was specified. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 735.694552] env[62460]: INFO nova.compute.manager [-] [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] Took 1.02 seconds to deallocate network for instance. [ 735.697758] env[62460]: DEBUG nova.compute.claims [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 735.697758] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 735.733231] env[62460]: DEBUG oslo_concurrency.lockutils [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] Releasing lock "refresh_cache-c4c5abc4-a5ed-4337-a930-d37d579819f8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 735.733679] env[62460]: DEBUG nova.compute.manager [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 735.733872] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 735.734207] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-deccdc7f-232a-4689-a0f8-c51e6657040f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.743970] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b2bb706-96ec-4eba-8373-ee373bbe40fb {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.769412] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c4c5abc4-a5ed-4337-a930-d37d579819f8 could not be found. [ 735.769647] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 735.769832] env[62460]: INFO nova.compute.manager [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] Took 0.04 seconds to destroy the instance on the hypervisor. [ 735.770098] env[62460]: DEBUG oslo.service.loopingcall [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 735.770334] env[62460]: DEBUG nova.compute.manager [-] [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 735.770430] env[62460]: DEBUG nova.network.neutron [-] [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 735.785095] env[62460]: DEBUG nova.network.neutron [-] [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 736.067034] env[62460]: DEBUG nova.compute.manager [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 4ed90d16-81a6-4dbd-8936-0e137151171f] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 736.281065] env[62460]: DEBUG nova.compute.manager [req-00ea2395-4790-411b-b4d9-b4073c301c7b req-5d811bca-2c81-4019-abd6-988ea3001a2c service nova] [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] Received event network-changed-6fc7ca35-5bb9-4d46-9356-ceb957531a76 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 736.281188] env[62460]: DEBUG nova.compute.manager [req-00ea2395-4790-411b-b4d9-b4073c301c7b req-5d811bca-2c81-4019-abd6-988ea3001a2c service nova] [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] Refreshing instance network info cache due to event network-changed-6fc7ca35-5bb9-4d46-9356-ceb957531a76. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 736.281398] env[62460]: DEBUG oslo_concurrency.lockutils [req-00ea2395-4790-411b-b4d9-b4073c301c7b req-5d811bca-2c81-4019-abd6-988ea3001a2c service nova] Acquiring lock "refresh_cache-c4c5abc4-a5ed-4337-a930-d37d579819f8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 736.281556] env[62460]: DEBUG oslo_concurrency.lockutils [req-00ea2395-4790-411b-b4d9-b4073c301c7b req-5d811bca-2c81-4019-abd6-988ea3001a2c service nova] Acquired lock "refresh_cache-c4c5abc4-a5ed-4337-a930-d37d579819f8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.281757] env[62460]: DEBUG nova.network.neutron [req-00ea2395-4790-411b-b4d9-b4073c301c7b req-5d811bca-2c81-4019-abd6-988ea3001a2c service nova] [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] Refreshing network info cache for port 6fc7ca35-5bb9-4d46-9356-ceb957531a76 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 736.287695] env[62460]: DEBUG nova.network.neutron [-] [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 736.368520] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9bbc135-986f-4a19-b856-55796add482d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.375932] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80faaa32-12be-492b-80e2-068d4d09c883 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.406044] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47ab0103-dd09-4f9c-9da4-f341a7429a0f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.413404] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ae0ede6-aebe-41ad-a29b-a6f8b0698121 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.426443] env[62460]: DEBUG nova.compute.provider_tree [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 736.791195] env[62460]: INFO nova.compute.manager [-] [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] Took 1.02 seconds to deallocate network for instance. [ 736.794947] env[62460]: DEBUG nova.compute.claims [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 736.795198] env[62460]: DEBUG oslo_concurrency.lockutils [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 736.801784] env[62460]: DEBUG nova.network.neutron [req-00ea2395-4790-411b-b4d9-b4073c301c7b req-5d811bca-2c81-4019-abd6-988ea3001a2c service nova] [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 736.891154] env[62460]: DEBUG nova.network.neutron [req-00ea2395-4790-411b-b4d9-b4073c301c7b req-5d811bca-2c81-4019-abd6-988ea3001a2c service nova] [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 736.929349] env[62460]: DEBUG nova.scheduler.client.report [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 737.077313] env[62460]: DEBUG nova.compute.manager [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 4ed90d16-81a6-4dbd-8936-0e137151171f] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 737.101420] env[62460]: DEBUG nova.virt.hardware [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 737.101570] env[62460]: DEBUG nova.virt.hardware [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 737.102147] env[62460]: DEBUG nova.virt.hardware [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 737.102147] env[62460]: DEBUG nova.virt.hardware [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 737.102147] env[62460]: DEBUG nova.virt.hardware [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 737.102334] env[62460]: DEBUG nova.virt.hardware [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 737.102453] env[62460]: DEBUG nova.virt.hardware [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 737.102617] env[62460]: DEBUG nova.virt.hardware [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 737.102786] env[62460]: DEBUG nova.virt.hardware [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 737.103058] env[62460]: DEBUG nova.virt.hardware [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 737.103348] env[62460]: DEBUG nova.virt.hardware [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 737.104392] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87132b70-5934-4546-b42e-a8cc40f9d8cf {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.112514] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e4bf6b4-27a4-43bb-8e31-1b4ad80e550c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.125714] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 4ed90d16-81a6-4dbd-8936-0e137151171f] Instance VIF info [] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 737.131151] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Creating folder: Project (8abc6846d0654fcdaa101e631b8b523e). Parent ref: group-v281134. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 737.131412] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-341e879d-618b-40d3-829f-b984703735bb {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.141340] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Created folder: Project (8abc6846d0654fcdaa101e631b8b523e) in parent group-v281134. [ 737.141480] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Creating folder: Instances. Parent ref: group-v281150. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 737.141693] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d71f3db4-02c9-463c-83f1-0a77320c0497 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.150503] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Created folder: Instances in parent group-v281150. [ 737.150725] env[62460]: DEBUG oslo.service.loopingcall [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 737.151188] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4ed90d16-81a6-4dbd-8936-0e137151171f] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 737.151188] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0bd38374-b3c2-4516-86e2-c066803f1df9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.176582] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 737.176582] env[62460]: value = "task-1313451" [ 737.176582] env[62460]: _type = "Task" [ 737.176582] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.186994] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313451, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.394236] env[62460]: DEBUG oslo_concurrency.lockutils [req-00ea2395-4790-411b-b4d9-b4073c301c7b req-5d811bca-2c81-4019-abd6-988ea3001a2c service nova] Releasing lock "refresh_cache-c4c5abc4-a5ed-4337-a930-d37d579819f8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 737.394998] env[62460]: DEBUG nova.compute.manager [req-00ea2395-4790-411b-b4d9-b4073c301c7b req-5d811bca-2c81-4019-abd6-988ea3001a2c service nova] [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] Received event network-vif-deleted-6fc7ca35-5bb9-4d46-9356-ceb957531a76 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 737.434651] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.381s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 737.435217] env[62460]: DEBUG nova.compute.manager [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 737.438878] env[62460]: DEBUG oslo_concurrency.lockutils [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.901s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 737.440289] env[62460]: INFO nova.compute.claims [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 737.687387] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313451, 'name': CreateVM_Task, 'duration_secs': 0.26386} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.687693] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4ed90d16-81a6-4dbd-8936-0e137151171f] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 737.688361] env[62460]: DEBUG oslo_concurrency.lockutils [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 737.688653] env[62460]: DEBUG oslo_concurrency.lockutils [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.689152] env[62460]: DEBUG oslo_concurrency.lockutils [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 737.689526] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9ebc95d2-1a57-4bd9-bcde-ffecda702571 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.696234] env[62460]: DEBUG oslo_vmware.api [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Waiting for the task: (returnval){ [ 737.696234] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52481e2f-a915-0c8f-af55-095d22bb1da6" [ 737.696234] env[62460]: _type = "Task" [ 737.696234] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.706975] env[62460]: DEBUG oslo_vmware.api [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52481e2f-a915-0c8f-af55-095d22bb1da6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.945361] env[62460]: DEBUG nova.compute.utils [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 737.946914] env[62460]: DEBUG nova.compute.manager [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Not allocating networking since 'none' was specified. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 738.206696] env[62460]: DEBUG oslo_vmware.api [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52481e2f-a915-0c8f-af55-095d22bb1da6, 'name': SearchDatastore_Task, 'duration_secs': 0.014299} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.206995] env[62460]: DEBUG oslo_concurrency.lockutils [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 738.207239] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 4ed90d16-81a6-4dbd-8936-0e137151171f] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 738.207473] env[62460]: DEBUG oslo_concurrency.lockutils [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.207619] env[62460]: DEBUG oslo_concurrency.lockutils [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.207799] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 738.208069] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-640b4c78-1cad-42b3-a215-c9e7b5753d08 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.216581] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 738.216714] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 738.217414] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b0c39fcf-d736-4e92-97b1-bcefd359f243 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.222913] env[62460]: DEBUG oslo_vmware.api [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Waiting for the task: (returnval){ [ 738.222913] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52682bb0-7dab-4af4-578c-17ceb098d5c2" [ 738.222913] env[62460]: _type = "Task" [ 738.222913] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.229955] env[62460]: DEBUG oslo_vmware.api [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52682bb0-7dab-4af4-578c-17ceb098d5c2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.450474] env[62460]: DEBUG nova.compute.manager [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 738.732527] env[62460]: DEBUG oslo_vmware.api [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52682bb0-7dab-4af4-578c-17ceb098d5c2, 'name': SearchDatastore_Task, 'duration_secs': 0.009029} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.735819] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a8dcc7e9-4f12-42f3-9ee7-fcb668310993 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.741070] env[62460]: DEBUG oslo_vmware.api [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Waiting for the task: (returnval){ [ 738.741070] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52fe9162-4b82-3b0b-9314-b7eb1d40bef4" [ 738.741070] env[62460]: _type = "Task" [ 738.741070] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.750506] env[62460]: DEBUG oslo_vmware.api [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52fe9162-4b82-3b0b-9314-b7eb1d40bef4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.760267] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a76932f7-b6c9-45a8-ae20-46f632440ca8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.766816] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4882efa0-49e7-4047-a655-7930e1a22ef8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.797230] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5692970-2d76-45e8-b737-bba772493d55 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.804522] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-873a8090-e899-4b55-af47-4d70610c5e51 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.817805] env[62460]: DEBUG nova.compute.provider_tree [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 739.251893] env[62460]: DEBUG oslo_vmware.api [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52fe9162-4b82-3b0b-9314-b7eb1d40bef4, 'name': SearchDatastore_Task, 'duration_secs': 0.009948} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.252248] env[62460]: DEBUG oslo_concurrency.lockutils [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.252529] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 4ed90d16-81a6-4dbd-8936-0e137151171f/4ed90d16-81a6-4dbd-8936-0e137151171f.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 739.252785] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-931cf064-1721-4be5-a4e5-0dab4879f0a8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.260074] env[62460]: DEBUG oslo_vmware.api [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Waiting for the task: (returnval){ [ 739.260074] env[62460]: value = "task-1313452" [ 739.260074] env[62460]: _type = "Task" [ 739.260074] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.267194] env[62460]: DEBUG oslo_vmware.api [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': task-1313452, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.325468] env[62460]: DEBUG nova.scheduler.client.report [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 739.462512] env[62460]: DEBUG nova.compute.manager [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 739.487868] env[62460]: DEBUG nova.virt.hardware [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 739.488204] env[62460]: DEBUG nova.virt.hardware [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 739.488385] env[62460]: DEBUG nova.virt.hardware [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 739.488648] env[62460]: DEBUG nova.virt.hardware [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 739.488842] env[62460]: DEBUG nova.virt.hardware [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 739.489055] env[62460]: DEBUG nova.virt.hardware [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 739.489352] env[62460]: DEBUG nova.virt.hardware [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 739.489553] env[62460]: DEBUG nova.virt.hardware [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 739.489734] env[62460]: DEBUG nova.virt.hardware [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 739.489931] env[62460]: DEBUG nova.virt.hardware [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 739.490140] env[62460]: DEBUG nova.virt.hardware [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 739.491173] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faafa065-60ce-4166-94d6-9900df0db42c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.501034] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2b8dbf5-5cc2-42de-805f-0c5d04f69df3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.515956] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Instance VIF info [] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 739.522078] env[62460]: DEBUG oslo.service.loopingcall [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 739.523228] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 739.523365] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e3443089-7056-40ff-a20b-d15b767cb1a7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.541822] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 739.541822] env[62460]: value = "task-1313453" [ 739.541822] env[62460]: _type = "Task" [ 739.541822] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.550980] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313453, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.769797] env[62460]: DEBUG oslo_vmware.api [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': task-1313452, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.464107} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.770090] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 4ed90d16-81a6-4dbd-8936-0e137151171f/4ed90d16-81a6-4dbd-8936-0e137151171f.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 739.770090] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 4ed90d16-81a6-4dbd-8936-0e137151171f] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 739.770294] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9c1e68d4-52cd-400d-aa54-88f3fb450292 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.777361] env[62460]: DEBUG oslo_vmware.api [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Waiting for the task: (returnval){ [ 739.777361] env[62460]: value = "task-1313454" [ 739.777361] env[62460]: _type = "Task" [ 739.777361] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.784711] env[62460]: DEBUG oslo_vmware.api [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': task-1313454, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.831836] env[62460]: DEBUG oslo_concurrency.lockutils [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.393s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 739.832430] env[62460]: DEBUG nova.compute.manager [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 739.835098] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 16.575s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.051639] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313453, 'name': CreateVM_Task, 'duration_secs': 0.302492} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.051809] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 740.052317] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 740.052415] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.052707] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 740.052952] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-644c39fa-1618-47fe-96ed-02eeb8df65d8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.058292] env[62460]: DEBUG oslo_vmware.api [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Waiting for the task: (returnval){ [ 740.058292] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]520a046c-96f1-096f-260a-645abbc52ab6" [ 740.058292] env[62460]: _type = "Task" [ 740.058292] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.065977] env[62460]: DEBUG oslo_vmware.api [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]520a046c-96f1-096f-260a-645abbc52ab6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.286266] env[62460]: DEBUG oslo_vmware.api [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': task-1313454, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066216} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.286549] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 4ed90d16-81a6-4dbd-8936-0e137151171f] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 740.287262] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d98a158-43ee-4790-a9f6-1af0d374d2d0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.306420] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 4ed90d16-81a6-4dbd-8936-0e137151171f] Reconfiguring VM instance instance-00000030 to attach disk [datastore1] 4ed90d16-81a6-4dbd-8936-0e137151171f/4ed90d16-81a6-4dbd-8936-0e137151171f.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 740.306664] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d99487e1-08d4-437e-be4a-e370b13b9ebd {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.325319] env[62460]: DEBUG oslo_vmware.api [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Waiting for the task: (returnval){ [ 740.325319] env[62460]: value = "task-1313455" [ 740.325319] env[62460]: _type = "Task" [ 740.325319] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.332607] env[62460]: DEBUG oslo_vmware.api [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': task-1313455, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.337831] env[62460]: DEBUG nova.compute.utils [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 740.342467] env[62460]: DEBUG nova.compute.manager [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 740.342581] env[62460]: DEBUG nova.network.neutron [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 740.394286] env[62460]: DEBUG nova.policy [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '632d7ca8cb4243f990d741a3f70f422a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '10ac54db1ef54e249a077bbb3dde4242', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 740.568609] env[62460]: DEBUG oslo_vmware.api [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]520a046c-96f1-096f-260a-645abbc52ab6, 'name': SearchDatastore_Task, 'duration_secs': 0.009413} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.569986] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.569986] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 740.569986] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 740.569986] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.570171] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 740.570171] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-99a1e126-c64a-43e2-84b0-528ea8415927 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.577638] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 740.577822] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 740.578552] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d3de2c7e-4b12-483f-810b-c70a29894016 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.584238] env[62460]: DEBUG oslo_vmware.api [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Waiting for the task: (returnval){ [ 740.584238] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52e36e78-d745-6505-cf6b-4d05ec1d8db1" [ 740.584238] env[62460]: _type = "Task" [ 740.584238] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.592574] env[62460]: DEBUG oslo_vmware.api [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52e36e78-d745-6505-cf6b-4d05ec1d8db1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.657560] env[62460]: DEBUG nova.network.neutron [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] Successfully created port: a9889184-6641-40e8-a98a-0942457eed29 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 740.838047] env[62460]: DEBUG oslo_vmware.api [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': task-1313455, 'name': ReconfigVM_Task, 'duration_secs': 0.300752} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.838047] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 4ed90d16-81a6-4dbd-8936-0e137151171f] Reconfigured VM instance instance-00000030 to attach disk [datastore1] 4ed90d16-81a6-4dbd-8936-0e137151171f/4ed90d16-81a6-4dbd-8936-0e137151171f.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 740.838047] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-aa2fd709-0310-457c-a6ee-c1a772f9d424 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.842769] env[62460]: DEBUG nova.compute.manager [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 740.845656] env[62460]: DEBUG oslo_vmware.api [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Waiting for the task: (returnval){ [ 740.845656] env[62460]: value = "task-1313456" [ 740.845656] env[62460]: _type = "Task" [ 740.845656] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.857843] env[62460]: DEBUG oslo_vmware.api [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': task-1313456, 'name': Rename_Task} progress is 6%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.866784] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance ec2297bb-1d29-4ea9-90f7-f4f39716c103 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 740.867015] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 0e00ca48-f9e1-4999-aad5-d7965f7ddc28 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 740.867200] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 740.867379] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance c4c5abc4-a5ed-4337-a930-d37d579819f8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 740.867549] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 4ed90d16-81a6-4dbd-8936-0e137151171f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 740.867698] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 5214f4a1-3e28-41bf-88d9-161511385e1b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 740.867858] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance e472b96a-5d82-4fb9-8d28-0901b71dc783 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 741.096637] env[62460]: DEBUG oslo_vmware.api [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52e36e78-d745-6505-cf6b-4d05ec1d8db1, 'name': SearchDatastore_Task, 'duration_secs': 0.0096} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.097677] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-272b2d3d-dd10-4b96-9fa9-4fc1f86f6ad6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.105824] env[62460]: DEBUG oslo_vmware.api [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Waiting for the task: (returnval){ [ 741.105824] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]5202b183-f836-e94b-b951-8f4c95521479" [ 741.105824] env[62460]: _type = "Task" [ 741.105824] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.117312] env[62460]: DEBUG oslo_vmware.api [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5202b183-f836-e94b-b951-8f4c95521479, 'name': SearchDatastore_Task, 'duration_secs': 0.008144} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.117758] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 741.117914] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 5214f4a1-3e28-41bf-88d9-161511385e1b/5214f4a1-3e28-41bf-88d9-161511385e1b.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 741.118207] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-78c6011e-7a2c-45ae-8548-982f7f18e9b8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.124937] env[62460]: DEBUG oslo_vmware.api [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Waiting for the task: (returnval){ [ 741.124937] env[62460]: value = "task-1313457" [ 741.124937] env[62460]: _type = "Task" [ 741.124937] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.133387] env[62460]: DEBUG oslo_vmware.api [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': task-1313457, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.362197] env[62460]: DEBUG oslo_vmware.api [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': task-1313456, 'name': Rename_Task, 'duration_secs': 0.134795} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.362197] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 4ed90d16-81a6-4dbd-8936-0e137151171f] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 741.362197] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cc2ed6fa-8fe9-4260-9f3a-a308f785563a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.370021] env[62460]: DEBUG oslo_vmware.api [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Waiting for the task: (returnval){ [ 741.370021] env[62460]: value = "task-1313458" [ 741.370021] env[62460]: _type = "Task" [ 741.370021] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.372588] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance be50b542-4c61-4da3-af4e-8a57e05a77ee has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 741.382244] env[62460]: DEBUG oslo_vmware.api [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': task-1313458, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.494674] env[62460]: DEBUG nova.compute.manager [req-316da5cf-87bc-497f-8f01-7d97fa52bd39 req-7ba9520f-bb23-47ab-b89b-81c0c71f64cc service nova] [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] Received event network-changed-a9889184-6641-40e8-a98a-0942457eed29 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 741.494891] env[62460]: DEBUG nova.compute.manager [req-316da5cf-87bc-497f-8f01-7d97fa52bd39 req-7ba9520f-bb23-47ab-b89b-81c0c71f64cc service nova] [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] Refreshing instance network info cache due to event network-changed-a9889184-6641-40e8-a98a-0942457eed29. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 741.495578] env[62460]: DEBUG oslo_concurrency.lockutils [req-316da5cf-87bc-497f-8f01-7d97fa52bd39 req-7ba9520f-bb23-47ab-b89b-81c0c71f64cc service nova] Acquiring lock "refresh_cache-e472b96a-5d82-4fb9-8d28-0901b71dc783" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 741.495758] env[62460]: DEBUG oslo_concurrency.lockutils [req-316da5cf-87bc-497f-8f01-7d97fa52bd39 req-7ba9520f-bb23-47ab-b89b-81c0c71f64cc service nova] Acquired lock "refresh_cache-e472b96a-5d82-4fb9-8d28-0901b71dc783" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.495937] env[62460]: DEBUG nova.network.neutron [req-316da5cf-87bc-497f-8f01-7d97fa52bd39 req-7ba9520f-bb23-47ab-b89b-81c0c71f64cc service nova] [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] Refreshing network info cache for port a9889184-6641-40e8-a98a-0942457eed29 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 741.634756] env[62460]: DEBUG oslo_vmware.api [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': task-1313457, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.471774} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.634972] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 5214f4a1-3e28-41bf-88d9-161511385e1b/5214f4a1-3e28-41bf-88d9-161511385e1b.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 741.635256] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 741.635510] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a84317a9-a153-4091-881f-8584a420f9ac {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.641312] env[62460]: DEBUG oslo_vmware.api [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Waiting for the task: (returnval){ [ 741.641312] env[62460]: value = "task-1313459" [ 741.641312] env[62460]: _type = "Task" [ 741.641312] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.649075] env[62460]: DEBUG oslo_vmware.api [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': task-1313459, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.688258] env[62460]: ERROR nova.compute.manager [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a9889184-6641-40e8-a98a-0942457eed29, please check neutron logs for more information. [ 741.688258] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 741.688258] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 741.688258] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 741.688258] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 741.688258] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 741.688258] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 741.688258] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 741.688258] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 741.688258] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 741.688258] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 741.688258] env[62460]: ERROR nova.compute.manager raise self.value [ 741.688258] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 741.688258] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 741.688258] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 741.688258] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 741.688836] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 741.688836] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 741.688836] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a9889184-6641-40e8-a98a-0942457eed29, please check neutron logs for more information. [ 741.688836] env[62460]: ERROR nova.compute.manager [ 741.688836] env[62460]: Traceback (most recent call last): [ 741.688836] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 741.688836] env[62460]: listener.cb(fileno) [ 741.688836] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 741.688836] env[62460]: result = function(*args, **kwargs) [ 741.688836] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 741.688836] env[62460]: return func(*args, **kwargs) [ 741.688836] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 741.688836] env[62460]: raise e [ 741.688836] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 741.688836] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 741.688836] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 741.688836] env[62460]: created_port_ids = self._update_ports_for_instance( [ 741.688836] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 741.688836] env[62460]: with excutils.save_and_reraise_exception(): [ 741.688836] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 741.688836] env[62460]: self.force_reraise() [ 741.688836] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 741.688836] env[62460]: raise self.value [ 741.688836] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 741.688836] env[62460]: updated_port = self._update_port( [ 741.688836] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 741.688836] env[62460]: _ensure_no_port_binding_failure(port) [ 741.688836] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 741.688836] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 741.689781] env[62460]: nova.exception.PortBindingFailed: Binding failed for port a9889184-6641-40e8-a98a-0942457eed29, please check neutron logs for more information. [ 741.689781] env[62460]: Removing descriptor: 18 [ 741.859021] env[62460]: DEBUG nova.compute.manager [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 741.880871] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 741.882699] env[62460]: DEBUG oslo_vmware.api [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': task-1313458, 'name': PowerOnVM_Task, 'duration_secs': 0.447289} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.885017] env[62460]: DEBUG nova.virt.hardware [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 741.885261] env[62460]: DEBUG nova.virt.hardware [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 741.885446] env[62460]: DEBUG nova.virt.hardware [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 741.885710] env[62460]: DEBUG nova.virt.hardware [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 741.885863] env[62460]: DEBUG nova.virt.hardware [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 741.886044] env[62460]: DEBUG nova.virt.hardware [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 741.886330] env[62460]: DEBUG nova.virt.hardware [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 741.886497] env[62460]: DEBUG nova.virt.hardware [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 741.886670] env[62460]: DEBUG nova.virt.hardware [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 741.886831] env[62460]: DEBUG nova.virt.hardware [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 741.887015] env[62460]: DEBUG nova.virt.hardware [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 741.887300] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 4ed90d16-81a6-4dbd-8936-0e137151171f] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 741.887493] env[62460]: INFO nova.compute.manager [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 4ed90d16-81a6-4dbd-8936-0e137151171f] Took 4.81 seconds to spawn the instance on the hypervisor. [ 741.887666] env[62460]: DEBUG nova.compute.manager [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 4ed90d16-81a6-4dbd-8936-0e137151171f] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 741.888429] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c43f2aa1-0a03-420f-899e-2a6c44af24e3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.891437] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98186e0a-d84c-4d72-92a5-0c8aefd8fd2e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.900852] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04cf2a22-ff18-4231-a16a-df490defde9a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.917275] env[62460]: ERROR nova.compute.manager [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a9889184-6641-40e8-a98a-0942457eed29, please check neutron logs for more information. [ 741.917275] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] Traceback (most recent call last): [ 741.917275] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 741.917275] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] yield resources [ 741.917275] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 741.917275] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] self.driver.spawn(context, instance, image_meta, [ 741.917275] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 741.917275] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] self._vmops.spawn(context, instance, image_meta, injected_files, [ 741.917275] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 741.917275] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] vm_ref = self.build_virtual_machine(instance, [ 741.917275] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 741.917909] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] vif_infos = vmwarevif.get_vif_info(self._session, [ 741.917909] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 741.917909] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] for vif in network_info: [ 741.917909] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 741.917909] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] return self._sync_wrapper(fn, *args, **kwargs) [ 741.917909] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 741.917909] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] self.wait() [ 741.917909] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 741.917909] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] self[:] = self._gt.wait() [ 741.917909] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 741.917909] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] return self._exit_event.wait() [ 741.917909] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 741.917909] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] current.throw(*self._exc) [ 741.918336] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 741.918336] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] result = function(*args, **kwargs) [ 741.918336] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 741.918336] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] return func(*args, **kwargs) [ 741.918336] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 741.918336] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] raise e [ 741.918336] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 741.918336] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] nwinfo = self.network_api.allocate_for_instance( [ 741.918336] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 741.918336] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] created_port_ids = self._update_ports_for_instance( [ 741.918336] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 741.918336] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] with excutils.save_and_reraise_exception(): [ 741.918336] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 741.918748] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] self.force_reraise() [ 741.918748] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 741.918748] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] raise self.value [ 741.918748] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 741.918748] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] updated_port = self._update_port( [ 741.918748] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 741.918748] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] _ensure_no_port_binding_failure(port) [ 741.918748] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 741.918748] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] raise exception.PortBindingFailed(port_id=port['id']) [ 741.918748] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] nova.exception.PortBindingFailed: Binding failed for port a9889184-6641-40e8-a98a-0942457eed29, please check neutron logs for more information. [ 741.918748] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] [ 741.918748] env[62460]: INFO nova.compute.manager [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] Terminating instance [ 741.919358] env[62460]: DEBUG oslo_concurrency.lockutils [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "refresh_cache-e472b96a-5d82-4fb9-8d28-0901b71dc783" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 742.011873] env[62460]: DEBUG nova.network.neutron [req-316da5cf-87bc-497f-8f01-7d97fa52bd39 req-7ba9520f-bb23-47ab-b89b-81c0c71f64cc service nova] [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 742.095859] env[62460]: DEBUG nova.network.neutron [req-316da5cf-87bc-497f-8f01-7d97fa52bd39 req-7ba9520f-bb23-47ab-b89b-81c0c71f64cc service nova] [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.151724] env[62460]: DEBUG oslo_vmware.api [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': task-1313459, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063848} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.152011] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 742.152973] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9817635d-7615-4568-b81d-0970f8fee4d6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.172785] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Reconfiguring VM instance instance-00000031 to attach disk [datastore1] 5214f4a1-3e28-41bf-88d9-161511385e1b/5214f4a1-3e28-41bf-88d9-161511385e1b.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 742.173467] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6352224a-6639-450c-97ea-40194ae3599d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.193526] env[62460]: DEBUG oslo_vmware.api [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Waiting for the task: (returnval){ [ 742.193526] env[62460]: value = "task-1313460" [ 742.193526] env[62460]: _type = "Task" [ 742.193526] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.205514] env[62460]: DEBUG oslo_vmware.api [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': task-1313460, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.384279] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 559757ef-ee92-4b88-8631-6f743fb88bc6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 742.413902] env[62460]: INFO nova.compute.manager [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 4ed90d16-81a6-4dbd-8936-0e137151171f] Took 24.80 seconds to build instance. [ 742.600034] env[62460]: DEBUG oslo_concurrency.lockutils [req-316da5cf-87bc-497f-8f01-7d97fa52bd39 req-7ba9520f-bb23-47ab-b89b-81c0c71f64cc service nova] Releasing lock "refresh_cache-e472b96a-5d82-4fb9-8d28-0901b71dc783" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.600549] env[62460]: DEBUG oslo_concurrency.lockutils [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquired lock "refresh_cache-e472b96a-5d82-4fb9-8d28-0901b71dc783" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.600781] env[62460]: DEBUG nova.network.neutron [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 742.703478] env[62460]: DEBUG oslo_vmware.api [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': task-1313460, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.887808] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 588e5489-2f61-473a-90e9-883eb8b163c8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 742.915871] env[62460]: DEBUG oslo_concurrency.lockutils [None req-78310499-dd33-418a-bdcc-ae7cc41b8ae6 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Lock "4ed90d16-81a6-4dbd-8936-0e137151171f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 109.749s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 743.118309] env[62460]: DEBUG nova.network.neutron [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 743.192447] env[62460]: DEBUG nova.network.neutron [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.205290] env[62460]: DEBUG oslo_vmware.api [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': task-1313460, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.391349] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 6c58f9fe-fb6d-4012-8b69-39e4134996b2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 743.418503] env[62460]: DEBUG nova.compute.manager [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 743.525218] env[62460]: DEBUG nova.compute.manager [req-2b689678-c4d6-43d2-ac39-f3e67f118aca req-44cb2d52-69f1-48f4-b2e6-d17a154bab5a service nova] [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] Received event network-vif-deleted-a9889184-6641-40e8-a98a-0942457eed29 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 743.695365] env[62460]: DEBUG oslo_concurrency.lockutils [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Releasing lock "refresh_cache-e472b96a-5d82-4fb9-8d28-0901b71dc783" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 743.695631] env[62460]: DEBUG nova.compute.manager [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 743.695827] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 743.696660] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6a6e86f1-7e0e-42bb-a961-4cf6e6bc5a8a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.707802] env[62460]: DEBUG oslo_vmware.api [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': task-1313460, 'name': ReconfigVM_Task, 'duration_secs': 1.050287} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.708955] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Reconfigured VM instance instance-00000031 to attach disk [datastore1] 5214f4a1-3e28-41bf-88d9-161511385e1b/5214f4a1-3e28-41bf-88d9-161511385e1b.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 743.709697] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-92ff8012-1893-4c9d-a3f0-a1106ac56f65 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.713570] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6456388-c1e0-44ef-9186-84ec293f0f27 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.727612] env[62460]: DEBUG oslo_vmware.api [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Waiting for the task: (returnval){ [ 743.727612] env[62460]: value = "task-1313461" [ 743.727612] env[62460]: _type = "Task" [ 743.727612] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.734915] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e472b96a-5d82-4fb9-8d28-0901b71dc783 could not be found. [ 743.735204] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 743.735327] env[62460]: INFO nova.compute.manager [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] Took 0.04 seconds to destroy the instance on the hypervisor. [ 743.735556] env[62460]: DEBUG oslo.service.loopingcall [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 743.738458] env[62460]: DEBUG nova.compute.manager [-] [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 743.738557] env[62460]: DEBUG nova.network.neutron [-] [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 743.740028] env[62460]: DEBUG oslo_vmware.api [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': task-1313461, 'name': Rename_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.754494] env[62460]: DEBUG nova.network.neutron [-] [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 743.895563] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance c7d51b39-b449-4af3-a4d0-c746983ded3e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 743.944563] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 744.237573] env[62460]: DEBUG oslo_vmware.api [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': task-1313461, 'name': Rename_Task, 'duration_secs': 0.115651} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.237966] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 744.238233] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b2502b00-f40d-46ac-ab55-941e62a3d9cf {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.244698] env[62460]: DEBUG oslo_vmware.api [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Waiting for the task: (returnval){ [ 744.244698] env[62460]: value = "task-1313462" [ 744.244698] env[62460]: _type = "Task" [ 744.244698] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.252480] env[62460]: DEBUG oslo_vmware.api [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': task-1313462, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.256858] env[62460]: DEBUG nova.network.neutron [-] [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.398920] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 51718896-f5bf-43a9-9396-1ac768737ba2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 744.755171] env[62460]: DEBUG oslo_vmware.api [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': task-1313462, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.759678] env[62460]: INFO nova.compute.manager [-] [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] Took 1.02 seconds to deallocate network for instance. [ 744.762228] env[62460]: DEBUG nova.compute.claims [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 744.762408] env[62460]: DEBUG oslo_concurrency.lockutils [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 744.902739] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 1f318a64-2c38-470b-8fae-4ba4543a5681 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 745.255645] env[62460]: DEBUG oslo_vmware.api [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': task-1313462, 'name': PowerOnVM_Task, 'duration_secs': 0.722668} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.255913] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 745.256163] env[62460]: INFO nova.compute.manager [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Took 5.79 seconds to spawn the instance on the hypervisor. [ 745.256365] env[62460]: DEBUG nova.compute.manager [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 745.257117] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d14e7fc-d554-4d3a-915b-6e6e159e174f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.407025] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 7e6ff902-4a04-43d5-9014-38c4ec88efc4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 745.774874] env[62460]: INFO nova.compute.manager [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Took 26.19 seconds to build instance. [ 745.910145] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance db09d1f5-88cc-4dc7-9a7b-5d53d09567fa has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 746.277884] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a6119075-eced-47d9-ac20-1d5116697167 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Lock "5214f4a1-3e28-41bf-88d9-161511385e1b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 112.842s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 746.414998] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 72e91f8e-0619-464c-b9bc-d6a14be42cb8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 746.779253] env[62460]: DEBUG nova.compute.manager [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 746.879659] env[62460]: INFO nova.compute.manager [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Rebuilding instance [ 746.915630] env[62460]: DEBUG nova.compute.manager [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 746.916515] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc98f6f2-e9e9-48f7-9012-6e6c2429f424 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.919553] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 3ada3516-3147-4566-a46a-1cb29cf880d0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 747.298457] env[62460]: DEBUG oslo_concurrency.lockutils [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 747.422462] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance fde12685-d3b9-46a0-8931-25b904d4f21e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 747.428412] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 747.428691] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-469a1e17-a19f-4d4b-a68c-411ca9b76bee {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.436600] env[62460]: DEBUG oslo_vmware.api [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Waiting for the task: (returnval){ [ 747.436600] env[62460]: value = "task-1313463" [ 747.436600] env[62460]: _type = "Task" [ 747.436600] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.446073] env[62460]: DEBUG oslo_vmware.api [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': task-1313463, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.925187] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 13890351-6091-4b4f-8484-1cdd0c8523b0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 747.948150] env[62460]: DEBUG oslo_vmware.api [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': task-1313463, 'name': PowerOffVM_Task, 'duration_secs': 0.160011} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.948430] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 747.948654] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 747.949422] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a072974-e673-42e9-aaad-7d8e0be986d1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.956645] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 747.956879] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a47eb362-0d07-4a16-a44e-ec8576c91f92 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.980519] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 747.980797] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Deleting contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 747.980970] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Deleting the datastore file [datastore1] 5214f4a1-3e28-41bf-88d9-161511385e1b {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 747.981243] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3723fdff-4eef-4672-9299-d3e261b98df9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.987628] env[62460]: DEBUG oslo_vmware.api [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Waiting for the task: (returnval){ [ 747.987628] env[62460]: value = "task-1313465" [ 747.987628] env[62460]: _type = "Task" [ 747.987628] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.995928] env[62460]: DEBUG oslo_vmware.api [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': task-1313465, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.428414] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance b76f37a0-91d0-4a01-9d95-9c6586081175 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 748.499973] env[62460]: DEBUG oslo_vmware.api [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': task-1313465, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.097483} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.500588] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 748.500819] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Deleted contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 748.501096] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 748.996374] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 5b6e8205-003e-49c0-a73d-be2e032a8272 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 749.498850] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 749.535158] env[62460]: DEBUG nova.virt.hardware [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 749.535653] env[62460]: DEBUG nova.virt.hardware [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 749.535653] env[62460]: DEBUG nova.virt.hardware [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 749.535881] env[62460]: DEBUG nova.virt.hardware [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 749.536083] env[62460]: DEBUG nova.virt.hardware [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 749.536283] env[62460]: DEBUG nova.virt.hardware [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 749.536532] env[62460]: DEBUG nova.virt.hardware [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 749.536706] env[62460]: DEBUG nova.virt.hardware [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 749.537161] env[62460]: DEBUG nova.virt.hardware [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 749.537161] env[62460]: DEBUG nova.virt.hardware [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 749.537293] env[62460]: DEBUG nova.virt.hardware [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 749.538089] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b74e670-fea0-4063-9caa-e40a9f5a4a6f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.546420] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0148b886-d34b-4af0-9095-51d7dd66f26f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.561783] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Instance VIF info [] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 749.567962] env[62460]: DEBUG oslo.service.loopingcall [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 749.567962] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 749.568115] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d2c1d268-8e46-4bcd-9de1-363d2ba08328 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.585049] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 749.585049] env[62460]: value = "task-1313466" [ 749.585049] env[62460]: _type = "Task" [ 749.585049] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.592690] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313466, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.002226] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 0269dc64-d2b1-43c5-bdf7-11d97e534819 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 750.094773] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313466, 'name': CreateVM_Task} progress is 99%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.508832] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 9d1a5830-f3c0-4d18-9338-16f7b6962c6a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 750.509140] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Total usable vcpus: 48, total allocated vcpus: 7 {{(pid=62460) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 750.509339] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1856MB phys_disk=200GB used_disk=7GB total_vcpus=48 used_vcpus=7 pci_stats=[] {{(pid=62460) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 750.595256] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313466, 'name': CreateVM_Task, 'duration_secs': 0.591701} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.597789] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 750.598765] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 750.598765] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.598765] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 750.600306] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-27685596-56c8-4e97-96ec-7bbbc30150b7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.603578] env[62460]: DEBUG oslo_vmware.api [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Waiting for the task: (returnval){ [ 750.603578] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52afffca-e36c-33d8-cfad-7f441b44bfe5" [ 750.603578] env[62460]: _type = "Task" [ 750.603578] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.613480] env[62460]: DEBUG oslo_vmware.api [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52afffca-e36c-33d8-cfad-7f441b44bfe5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.814500] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4a0d805-7a21-4db5-a4ab-414a729f5d1c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.821928] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d56f7d7-69c1-4b27-addf-d2feab21fd7f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.851878] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3585e43-784b-48e7-8678-e74a01a4e6b3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.858731] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e0451dc-99de-49e0-8a6f-62fb1624b050 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.871718] env[62460]: DEBUG nova.compute.provider_tree [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 751.113847] env[62460]: DEBUG oslo_vmware.api [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52afffca-e36c-33d8-cfad-7f441b44bfe5, 'name': SearchDatastore_Task, 'duration_secs': 0.00965} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.115059] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.115059] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 751.115059] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 751.115059] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.115280] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 751.115473] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0d0b6ac0-478c-4437-a68d-b1c07856f31c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.123128] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 751.123329] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 751.123995] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5b7405c7-c30c-4307-aa11-164a9965e684 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.129157] env[62460]: DEBUG oslo_vmware.api [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Waiting for the task: (returnval){ [ 751.129157] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]525b2458-c113-e58a-e394-8cb9a8d75eff" [ 751.129157] env[62460]: _type = "Task" [ 751.129157] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.136668] env[62460]: DEBUG oslo_vmware.api [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]525b2458-c113-e58a-e394-8cb9a8d75eff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.377054] env[62460]: DEBUG nova.scheduler.client.report [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 751.639967] env[62460]: DEBUG oslo_vmware.api [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]525b2458-c113-e58a-e394-8cb9a8d75eff, 'name': SearchDatastore_Task, 'duration_secs': 0.007802} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.640569] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dceb9dc3-5d86-4650-86b1-e072a5cdf933 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.645669] env[62460]: DEBUG oslo_vmware.api [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Waiting for the task: (returnval){ [ 751.645669] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52b62014-c21e-5dd0-be79-8c373c321eac" [ 751.645669] env[62460]: _type = "Task" [ 751.645669] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.653432] env[62460]: DEBUG oslo_vmware.api [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52b62014-c21e-5dd0-be79-8c373c321eac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.881496] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62460) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 751.881752] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 12.047s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 751.882041] env[62460]: DEBUG oslo_concurrency.lockutils [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.326s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.883582] env[62460]: INFO nova.compute.claims [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 752.155811] env[62460]: DEBUG oslo_vmware.api [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52b62014-c21e-5dd0-be79-8c373c321eac, 'name': SearchDatastore_Task, 'duration_secs': 0.008298} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.156095] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.156360] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] 5214f4a1-3e28-41bf-88d9-161511385e1b/5214f4a1-3e28-41bf-88d9-161511385e1b.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 752.156617] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c31dd818-966b-4931-914f-18eb5da61f13 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.164197] env[62460]: DEBUG oslo_vmware.api [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Waiting for the task: (returnval){ [ 752.164197] env[62460]: value = "task-1313467" [ 752.164197] env[62460]: _type = "Task" [ 752.164197] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.171814] env[62460]: DEBUG oslo_vmware.api [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': task-1313467, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.674334] env[62460]: DEBUG oslo_vmware.api [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': task-1313467, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.432463} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.674624] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] 5214f4a1-3e28-41bf-88d9-161511385e1b/5214f4a1-3e28-41bf-88d9-161511385e1b.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 752.674805] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 752.675056] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-28be339a-65a9-4d40-bfaa-35d99c4de28f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.681498] env[62460]: DEBUG oslo_vmware.api [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Waiting for the task: (returnval){ [ 752.681498] env[62460]: value = "task-1313468" [ 752.681498] env[62460]: _type = "Task" [ 752.681498] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.688939] env[62460]: DEBUG oslo_vmware.api [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': task-1313468, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.165791] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4afddc1-98df-411e-a66f-6aaf745d691a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.173458] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1463cda2-3016-4114-aca7-0d14e237b504 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.204675] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da9fb9f8-d1e6-4bd5-bc3c-6b092aa16031 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.212074] env[62460]: DEBUG oslo_vmware.api [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': task-1313468, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.05573} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.213631] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 753.214799] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6568e0a-f099-4efd-a30f-a03391989e43 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.217547] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-660b704a-7347-4416-af2b-d1ebd74c4385 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.238785] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Reconfiguring VM instance instance-00000031 to attach disk [datastore2] 5214f4a1-3e28-41bf-88d9-161511385e1b/5214f4a1-3e28-41bf-88d9-161511385e1b.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 753.246270] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-48866a58-1425-4c35-9e90-ebe8505e4c99 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.259988] env[62460]: DEBUG nova.compute.provider_tree [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 753.266374] env[62460]: DEBUG oslo_vmware.api [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Waiting for the task: (returnval){ [ 753.266374] env[62460]: value = "task-1313469" [ 753.266374] env[62460]: _type = "Task" [ 753.266374] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.274652] env[62460]: DEBUG oslo_vmware.api [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': task-1313469, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.763059] env[62460]: DEBUG nova.scheduler.client.report [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 753.776454] env[62460]: DEBUG oslo_vmware.api [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': task-1313469, 'name': ReconfigVM_Task, 'duration_secs': 0.256691} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.777300] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Reconfigured VM instance instance-00000031 to attach disk [datastore2] 5214f4a1-3e28-41bf-88d9-161511385e1b/5214f4a1-3e28-41bf-88d9-161511385e1b.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 753.777868] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bb8bac92-1700-4ae1-8169-632fab090524 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.784837] env[62460]: DEBUG oslo_vmware.api [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Waiting for the task: (returnval){ [ 753.784837] env[62460]: value = "task-1313470" [ 753.784837] env[62460]: _type = "Task" [ 753.784837] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.793007] env[62460]: DEBUG oslo_vmware.api [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': task-1313470, 'name': Rename_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.267423] env[62460]: DEBUG oslo_concurrency.lockutils [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.385s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 754.268126] env[62460]: DEBUG nova.compute.manager [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 754.270915] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.831s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 754.272361] env[62460]: INFO nova.compute.claims [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 754.295855] env[62460]: DEBUG oslo_vmware.api [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': task-1313470, 'name': Rename_Task, 'duration_secs': 0.177089} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.295855] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 754.296155] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a1b92731-1cdc-4f74-9671-3a9cb674a6c7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.302841] env[62460]: DEBUG oslo_vmware.api [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Waiting for the task: (returnval){ [ 754.302841] env[62460]: value = "task-1313471" [ 754.302841] env[62460]: _type = "Task" [ 754.302841] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.311565] env[62460]: DEBUG oslo_vmware.api [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': task-1313471, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.776595] env[62460]: DEBUG nova.compute.utils [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 754.779873] env[62460]: DEBUG nova.compute.manager [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 754.780065] env[62460]: DEBUG nova.network.neutron [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 754.813560] env[62460]: DEBUG oslo_vmware.api [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': task-1313471, 'name': PowerOnVM_Task, 'duration_secs': 0.395675} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.814318] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 754.814318] env[62460]: DEBUG nova.compute.manager [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 754.815009] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76f4ac19-9c5f-4d16-9cdf-b50d6b94bcc2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.840497] env[62460]: DEBUG nova.policy [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4d95c2a24f0841de8990410e24f325e9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '11491a5610734d96bac82035c46bd5d1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 755.141666] env[62460]: DEBUG nova.network.neutron [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] Successfully created port: 0df43069-277e-4946-8608-d75983f32608 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 755.287618] env[62460]: DEBUG nova.compute.manager [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 755.338156] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 755.623266] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09d79179-dd0b-4b94-af17-13182cb195c1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.631090] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76afcd5d-12c1-46ff-bcc1-e919aecb7941 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.660985] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7993a20b-0eb0-4d5f-a23e-e8e95b0f64a6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.668084] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0524634c-6472-4868-a42d-eee8b5a5e9da {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.681256] env[62460]: DEBUG nova.compute.provider_tree [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 755.934303] env[62460]: DEBUG nova.compute.manager [req-bad61583-bb7d-4c4e-bf32-d0db5a2d0d71 req-8ec2497b-422b-4d1a-817a-d3f30e8ed882 service nova] [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] Received event network-changed-0df43069-277e-4946-8608-d75983f32608 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 755.934609] env[62460]: DEBUG nova.compute.manager [req-bad61583-bb7d-4c4e-bf32-d0db5a2d0d71 req-8ec2497b-422b-4d1a-817a-d3f30e8ed882 service nova] [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] Refreshing instance network info cache due to event network-changed-0df43069-277e-4946-8608-d75983f32608. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 755.934788] env[62460]: DEBUG oslo_concurrency.lockutils [req-bad61583-bb7d-4c4e-bf32-d0db5a2d0d71 req-8ec2497b-422b-4d1a-817a-d3f30e8ed882 service nova] Acquiring lock "refresh_cache-be50b542-4c61-4da3-af4e-8a57e05a77ee" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 755.934840] env[62460]: DEBUG oslo_concurrency.lockutils [req-bad61583-bb7d-4c4e-bf32-d0db5a2d0d71 req-8ec2497b-422b-4d1a-817a-d3f30e8ed882 service nova] Acquired lock "refresh_cache-be50b542-4c61-4da3-af4e-8a57e05a77ee" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.934999] env[62460]: DEBUG nova.network.neutron [req-bad61583-bb7d-4c4e-bf32-d0db5a2d0d71 req-8ec2497b-422b-4d1a-817a-d3f30e8ed882 service nova] [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] Refreshing network info cache for port 0df43069-277e-4946-8608-d75983f32608 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 755.966106] env[62460]: DEBUG oslo_concurrency.lockutils [None req-59e9b2ee-6b25-41d0-8af5-fad3a3964ef2 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Acquiring lock "5214f4a1-3e28-41bf-88d9-161511385e1b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 755.966106] env[62460]: DEBUG oslo_concurrency.lockutils [None req-59e9b2ee-6b25-41d0-8af5-fad3a3964ef2 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Lock "5214f4a1-3e28-41bf-88d9-161511385e1b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 755.966106] env[62460]: DEBUG oslo_concurrency.lockutils [None req-59e9b2ee-6b25-41d0-8af5-fad3a3964ef2 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Acquiring lock "5214f4a1-3e28-41bf-88d9-161511385e1b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 755.966325] env[62460]: DEBUG oslo_concurrency.lockutils [None req-59e9b2ee-6b25-41d0-8af5-fad3a3964ef2 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Lock "5214f4a1-3e28-41bf-88d9-161511385e1b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 755.966325] env[62460]: DEBUG oslo_concurrency.lockutils [None req-59e9b2ee-6b25-41d0-8af5-fad3a3964ef2 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Lock "5214f4a1-3e28-41bf-88d9-161511385e1b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 755.968569] env[62460]: INFO nova.compute.manager [None req-59e9b2ee-6b25-41d0-8af5-fad3a3964ef2 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Terminating instance [ 755.970304] env[62460]: DEBUG oslo_concurrency.lockutils [None req-59e9b2ee-6b25-41d0-8af5-fad3a3964ef2 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Acquiring lock "refresh_cache-5214f4a1-3e28-41bf-88d9-161511385e1b" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 755.970411] env[62460]: DEBUG oslo_concurrency.lockutils [None req-59e9b2ee-6b25-41d0-8af5-fad3a3964ef2 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Acquired lock "refresh_cache-5214f4a1-3e28-41bf-88d9-161511385e1b" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.970578] env[62460]: DEBUG nova.network.neutron [None req-59e9b2ee-6b25-41d0-8af5-fad3a3964ef2 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 756.129304] env[62460]: ERROR nova.compute.manager [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0df43069-277e-4946-8608-d75983f32608, please check neutron logs for more information. [ 756.129304] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 756.129304] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 756.129304] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 756.129304] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 756.129304] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 756.129304] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 756.129304] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 756.129304] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 756.129304] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 756.129304] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 756.129304] env[62460]: ERROR nova.compute.manager raise self.value [ 756.129304] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 756.129304] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 756.129304] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 756.129304] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 756.129946] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 756.129946] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 756.129946] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0df43069-277e-4946-8608-d75983f32608, please check neutron logs for more information. [ 756.129946] env[62460]: ERROR nova.compute.manager [ 756.129946] env[62460]: Traceback (most recent call last): [ 756.129946] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 756.129946] env[62460]: listener.cb(fileno) [ 756.129946] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 756.129946] env[62460]: result = function(*args, **kwargs) [ 756.129946] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 756.129946] env[62460]: return func(*args, **kwargs) [ 756.129946] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 756.129946] env[62460]: raise e [ 756.129946] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 756.129946] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 756.129946] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 756.129946] env[62460]: created_port_ids = self._update_ports_for_instance( [ 756.129946] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 756.129946] env[62460]: with excutils.save_and_reraise_exception(): [ 756.129946] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 756.129946] env[62460]: self.force_reraise() [ 756.129946] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 756.129946] env[62460]: raise self.value [ 756.129946] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 756.129946] env[62460]: updated_port = self._update_port( [ 756.129946] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 756.129946] env[62460]: _ensure_no_port_binding_failure(port) [ 756.129946] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 756.129946] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 756.131095] env[62460]: nova.exception.PortBindingFailed: Binding failed for port 0df43069-277e-4946-8608-d75983f32608, please check neutron logs for more information. [ 756.131095] env[62460]: Removing descriptor: 18 [ 756.184872] env[62460]: DEBUG nova.scheduler.client.report [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 756.303037] env[62460]: DEBUG nova.compute.manager [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 756.327400] env[62460]: DEBUG nova.virt.hardware [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 756.327705] env[62460]: DEBUG nova.virt.hardware [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 756.327884] env[62460]: DEBUG nova.virt.hardware [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 756.328130] env[62460]: DEBUG nova.virt.hardware [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 756.328315] env[62460]: DEBUG nova.virt.hardware [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 756.328504] env[62460]: DEBUG nova.virt.hardware [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 756.328756] env[62460]: DEBUG nova.virt.hardware [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 756.328961] env[62460]: DEBUG nova.virt.hardware [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 756.329178] env[62460]: DEBUG nova.virt.hardware [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 756.329390] env[62460]: DEBUG nova.virt.hardware [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 756.329615] env[62460]: DEBUG nova.virt.hardware [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 756.330866] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-941f613b-f327-458a-b031-892393e03e77 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.339435] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1396cd72-d467-42a7-ba86-d313284c39ef {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.354203] env[62460]: ERROR nova.compute.manager [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0df43069-277e-4946-8608-d75983f32608, please check neutron logs for more information. [ 756.354203] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] Traceback (most recent call last): [ 756.354203] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 756.354203] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] yield resources [ 756.354203] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 756.354203] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] self.driver.spawn(context, instance, image_meta, [ 756.354203] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 756.354203] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] self._vmops.spawn(context, instance, image_meta, injected_files, [ 756.354203] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 756.354203] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] vm_ref = self.build_virtual_machine(instance, [ 756.354203] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 756.354667] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] vif_infos = vmwarevif.get_vif_info(self._session, [ 756.354667] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 756.354667] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] for vif in network_info: [ 756.354667] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 756.354667] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] return self._sync_wrapper(fn, *args, **kwargs) [ 756.354667] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 756.354667] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] self.wait() [ 756.354667] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 756.354667] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] self[:] = self._gt.wait() [ 756.354667] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 756.354667] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] return self._exit_event.wait() [ 756.354667] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 756.354667] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] current.throw(*self._exc) [ 756.355231] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 756.355231] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] result = function(*args, **kwargs) [ 756.355231] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 756.355231] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] return func(*args, **kwargs) [ 756.355231] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 756.355231] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] raise e [ 756.355231] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 756.355231] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] nwinfo = self.network_api.allocate_for_instance( [ 756.355231] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 756.355231] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] created_port_ids = self._update_ports_for_instance( [ 756.355231] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 756.355231] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] with excutils.save_and_reraise_exception(): [ 756.355231] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 756.355968] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] self.force_reraise() [ 756.355968] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 756.355968] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] raise self.value [ 756.355968] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 756.355968] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] updated_port = self._update_port( [ 756.355968] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 756.355968] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] _ensure_no_port_binding_failure(port) [ 756.355968] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 756.355968] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] raise exception.PortBindingFailed(port_id=port['id']) [ 756.355968] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] nova.exception.PortBindingFailed: Binding failed for port 0df43069-277e-4946-8608-d75983f32608, please check neutron logs for more information. [ 756.355968] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] [ 756.355968] env[62460]: INFO nova.compute.manager [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] Terminating instance [ 756.356552] env[62460]: DEBUG oslo_concurrency.lockutils [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Acquiring lock "refresh_cache-be50b542-4c61-4da3-af4e-8a57e05a77ee" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 756.453560] env[62460]: DEBUG nova.network.neutron [req-bad61583-bb7d-4c4e-bf32-d0db5a2d0d71 req-8ec2497b-422b-4d1a-817a-d3f30e8ed882 service nova] [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 756.487601] env[62460]: DEBUG nova.network.neutron [None req-59e9b2ee-6b25-41d0-8af5-fad3a3964ef2 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 756.496391] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 756.496391] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 756.537190] env[62460]: DEBUG nova.network.neutron [None req-59e9b2ee-6b25-41d0-8af5-fad3a3964ef2 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.555151] env[62460]: DEBUG nova.network.neutron [req-bad61583-bb7d-4c4e-bf32-d0db5a2d0d71 req-8ec2497b-422b-4d1a-817a-d3f30e8ed882 service nova] [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.692018] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.419s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 756.692018] env[62460]: DEBUG nova.compute.manager [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 756.693217] env[62460]: DEBUG oslo_concurrency.lockutils [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.346s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 757.003301] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 757.003301] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Starting heal instance info cache {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 757.003301] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Rebuilding the list of instances to heal {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 757.039908] env[62460]: DEBUG oslo_concurrency.lockutils [None req-59e9b2ee-6b25-41d0-8af5-fad3a3964ef2 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Releasing lock "refresh_cache-5214f4a1-3e28-41bf-88d9-161511385e1b" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 757.040368] env[62460]: DEBUG nova.compute.manager [None req-59e9b2ee-6b25-41d0-8af5-fad3a3964ef2 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 757.040567] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-59e9b2ee-6b25-41d0-8af5-fad3a3964ef2 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 757.041536] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e640ca2-f6c1-4de9-87c6-74d464755064 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.049513] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-59e9b2ee-6b25-41d0-8af5-fad3a3964ef2 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 757.049750] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3d814c08-07e3-412f-bb1e-d745ce1ca800 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.055903] env[62460]: DEBUG oslo_vmware.api [None req-59e9b2ee-6b25-41d0-8af5-fad3a3964ef2 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Waiting for the task: (returnval){ [ 757.055903] env[62460]: value = "task-1313472" [ 757.055903] env[62460]: _type = "Task" [ 757.055903] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.059428] env[62460]: DEBUG oslo_concurrency.lockutils [req-bad61583-bb7d-4c4e-bf32-d0db5a2d0d71 req-8ec2497b-422b-4d1a-817a-d3f30e8ed882 service nova] Releasing lock "refresh_cache-be50b542-4c61-4da3-af4e-8a57e05a77ee" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 757.059880] env[62460]: DEBUG oslo_concurrency.lockutils [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Acquired lock "refresh_cache-be50b542-4c61-4da3-af4e-8a57e05a77ee" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.060078] env[62460]: DEBUG nova.network.neutron [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 757.065789] env[62460]: DEBUG oslo_vmware.api [None req-59e9b2ee-6b25-41d0-8af5-fad3a3964ef2 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': task-1313472, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.198120] env[62460]: DEBUG nova.compute.utils [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 757.204148] env[62460]: DEBUG nova.compute.manager [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 757.204148] env[62460]: DEBUG nova.network.neutron [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 757.248063] env[62460]: DEBUG nova.policy [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e7460403518b4b399524dcdfbc07e338', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6ca24f5246c94585b07c3dc8e503b277', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 757.509903] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] Skipping network cache update for instance because it is Building. {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 757.509903] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] Skipping network cache update for instance because it is Building. {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 757.510048] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] Skipping network cache update for instance because it is Building. {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 757.510306] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Skipping network cache update for instance because it is being deleted. {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 757.510484] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] Skipping network cache update for instance because it is Building. {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 757.510631] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] Skipping network cache update for instance because it is Building. {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 757.510756] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] Skipping network cache update for instance because it is Building. {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 757.514214] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5cf1de7-3bcf-4053-b427-465f7b7b5180 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.524207] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f376eae-378c-487f-925b-d634e679f041 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.560775] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Acquiring lock "refresh_cache-4ed90d16-81a6-4dbd-8936-0e137151171f" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 757.560944] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Acquired lock "refresh_cache-4ed90d16-81a6-4dbd-8936-0e137151171f" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.561106] env[62460]: DEBUG nova.network.neutron [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 4ed90d16-81a6-4dbd-8936-0e137151171f] Forcefully refreshing network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 757.561263] env[62460]: DEBUG nova.objects.instance [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lazy-loading 'info_cache' on Instance uuid 4ed90d16-81a6-4dbd-8936-0e137151171f {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 757.565681] env[62460]: DEBUG nova.network.neutron [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] Successfully created port: 44c1e29f-674f-41c8-9ce8-737c88692e0b {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 757.570725] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a2f6980-b28c-446c-98bf-eb922dd10b99 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.579460] env[62460]: DEBUG oslo_vmware.api [None req-59e9b2ee-6b25-41d0-8af5-fad3a3964ef2 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': task-1313472, 'name': PowerOffVM_Task, 'duration_secs': 0.19524} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.581485] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-59e9b2ee-6b25-41d0-8af5-fad3a3964ef2 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 757.581662] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-59e9b2ee-6b25-41d0-8af5-fad3a3964ef2 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 757.581935] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5186eed0-665f-43fe-9276-6ef2bc6df5e1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.584285] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53175016-0c11-410b-9d7f-727dddecd3a9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.598392] env[62460]: DEBUG nova.compute.provider_tree [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 757.600279] env[62460]: DEBUG nova.network.neutron [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 757.613250] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-59e9b2ee-6b25-41d0-8af5-fad3a3964ef2 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 757.613250] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-59e9b2ee-6b25-41d0-8af5-fad3a3964ef2 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Deleting contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 757.613250] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-59e9b2ee-6b25-41d0-8af5-fad3a3964ef2 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Deleting the datastore file [datastore2] 5214f4a1-3e28-41bf-88d9-161511385e1b {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 757.613250] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8095457e-4bdc-4b6e-8b4e-7ce490ef7908 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.620254] env[62460]: DEBUG oslo_vmware.api [None req-59e9b2ee-6b25-41d0-8af5-fad3a3964ef2 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Waiting for the task: (returnval){ [ 757.620254] env[62460]: value = "task-1313474" [ 757.620254] env[62460]: _type = "Task" [ 757.620254] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.628462] env[62460]: DEBUG oslo_vmware.api [None req-59e9b2ee-6b25-41d0-8af5-fad3a3964ef2 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': task-1313474, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.682357] env[62460]: DEBUG nova.network.neutron [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 757.704304] env[62460]: DEBUG nova.compute.manager [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 757.961124] env[62460]: DEBUG nova.compute.manager [req-999fd942-c6e9-429d-ae46-10e344f9e6f9 req-6eae99bb-9744-4726-b620-8a0e9cfe6ab0 service nova] [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] Received event network-vif-deleted-0df43069-277e-4946-8608-d75983f32608 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 758.106017] env[62460]: DEBUG nova.scheduler.client.report [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 758.132257] env[62460]: DEBUG oslo_vmware.api [None req-59e9b2ee-6b25-41d0-8af5-fad3a3964ef2 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': task-1313474, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.096663} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.134333] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-59e9b2ee-6b25-41d0-8af5-fad3a3964ef2 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 758.134333] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-59e9b2ee-6b25-41d0-8af5-fad3a3964ef2 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Deleted contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 758.134333] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-59e9b2ee-6b25-41d0-8af5-fad3a3964ef2 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 758.134333] env[62460]: INFO nova.compute.manager [None req-59e9b2ee-6b25-41d0-8af5-fad3a3964ef2 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Took 1.09 seconds to destroy the instance on the hypervisor. [ 758.134333] env[62460]: DEBUG oslo.service.loopingcall [None req-59e9b2ee-6b25-41d0-8af5-fad3a3964ef2 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 758.134589] env[62460]: DEBUG nova.compute.manager [-] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 758.134589] env[62460]: DEBUG nova.network.neutron [-] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 758.149089] env[62460]: DEBUG nova.network.neutron [-] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 758.185457] env[62460]: DEBUG oslo_concurrency.lockutils [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Releasing lock "refresh_cache-be50b542-4c61-4da3-af4e-8a57e05a77ee" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 758.185916] env[62460]: DEBUG nova.compute.manager [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 758.186122] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 758.186413] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f5aac6f3-00d3-4b3f-939e-14a96277016e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.197088] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e2ed4ba-4a38-404a-942e-22f129d67399 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.218348] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance be50b542-4c61-4da3-af4e-8a57e05a77ee could not be found. [ 758.218568] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 758.218747] env[62460]: INFO nova.compute.manager [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] Took 0.03 seconds to destroy the instance on the hypervisor. [ 758.218989] env[62460]: DEBUG oslo.service.loopingcall [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 758.219218] env[62460]: DEBUG nova.compute.manager [-] [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 758.219314] env[62460]: DEBUG nova.network.neutron [-] [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 758.235474] env[62460]: DEBUG nova.network.neutron [-] [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 758.553946] env[62460]: ERROR nova.compute.manager [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 44c1e29f-674f-41c8-9ce8-737c88692e0b, please check neutron logs for more information. [ 758.553946] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 758.553946] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 758.553946] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 758.553946] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 758.553946] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 758.553946] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 758.553946] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 758.553946] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 758.553946] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 758.553946] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 758.553946] env[62460]: ERROR nova.compute.manager raise self.value [ 758.553946] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 758.553946] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 758.553946] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 758.553946] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 758.554591] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 758.554591] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 758.554591] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 44c1e29f-674f-41c8-9ce8-737c88692e0b, please check neutron logs for more information. [ 758.554591] env[62460]: ERROR nova.compute.manager [ 758.554591] env[62460]: Traceback (most recent call last): [ 758.554591] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 758.554591] env[62460]: listener.cb(fileno) [ 758.554591] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 758.554591] env[62460]: result = function(*args, **kwargs) [ 758.554591] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 758.554591] env[62460]: return func(*args, **kwargs) [ 758.554591] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 758.554591] env[62460]: raise e [ 758.554591] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 758.554591] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 758.554591] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 758.554591] env[62460]: created_port_ids = self._update_ports_for_instance( [ 758.554591] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 758.554591] env[62460]: with excutils.save_and_reraise_exception(): [ 758.554591] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 758.554591] env[62460]: self.force_reraise() [ 758.554591] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 758.554591] env[62460]: raise self.value [ 758.554591] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 758.554591] env[62460]: updated_port = self._update_port( [ 758.554591] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 758.554591] env[62460]: _ensure_no_port_binding_failure(port) [ 758.554591] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 758.554591] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 758.555602] env[62460]: nova.exception.PortBindingFailed: Binding failed for port 44c1e29f-674f-41c8-9ce8-737c88692e0b, please check neutron logs for more information. [ 758.555602] env[62460]: Removing descriptor: 18 [ 758.590711] env[62460]: DEBUG nova.network.neutron [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 4ed90d16-81a6-4dbd-8936-0e137151171f] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 758.610061] env[62460]: DEBUG oslo_concurrency.lockutils [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.917s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 758.610711] env[62460]: ERROR nova.compute.manager [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 75a13741-f486-4db0-be5b-054880a3cf89, please check neutron logs for more information. [ 758.610711] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] Traceback (most recent call last): [ 758.610711] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 758.610711] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] self.driver.spawn(context, instance, image_meta, [ 758.610711] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 758.610711] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] self._vmops.spawn(context, instance, image_meta, injected_files, [ 758.610711] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 758.610711] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] vm_ref = self.build_virtual_machine(instance, [ 758.610711] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 758.610711] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] vif_infos = vmwarevif.get_vif_info(self._session, [ 758.610711] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 758.611215] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] for vif in network_info: [ 758.611215] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 758.611215] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] return self._sync_wrapper(fn, *args, **kwargs) [ 758.611215] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 758.611215] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] self.wait() [ 758.611215] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 758.611215] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] self[:] = self._gt.wait() [ 758.611215] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 758.611215] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] return self._exit_event.wait() [ 758.611215] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 758.611215] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] result = hub.switch() [ 758.611215] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 758.611215] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] return self.greenlet.switch() [ 758.611890] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 758.611890] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] result = function(*args, **kwargs) [ 758.611890] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 758.611890] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] return func(*args, **kwargs) [ 758.611890] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 758.611890] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] raise e [ 758.611890] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 758.611890] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] nwinfo = self.network_api.allocate_for_instance( [ 758.611890] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 758.611890] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] created_port_ids = self._update_ports_for_instance( [ 758.611890] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 758.611890] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] with excutils.save_and_reraise_exception(): [ 758.611890] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 758.612942] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] self.force_reraise() [ 758.612942] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 758.612942] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] raise self.value [ 758.612942] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 758.612942] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] updated_port = self._update_port( [ 758.612942] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 758.612942] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] _ensure_no_port_binding_failure(port) [ 758.612942] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 758.612942] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] raise exception.PortBindingFailed(port_id=port['id']) [ 758.612942] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] nova.exception.PortBindingFailed: Binding failed for port 75a13741-f486-4db0-be5b-054880a3cf89, please check neutron logs for more information. [ 758.612942] env[62460]: ERROR nova.compute.manager [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] [ 758.613545] env[62460]: DEBUG nova.compute.utils [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] Binding failed for port 75a13741-f486-4db0-be5b-054880a3cf89, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 758.613545] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.252s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 758.615660] env[62460]: DEBUG nova.compute.manager [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] Build of instance ec2297bb-1d29-4ea9-90f7-f4f39716c103 was re-scheduled: Binding failed for port 75a13741-f486-4db0-be5b-054880a3cf89, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 758.616114] env[62460]: DEBUG nova.compute.manager [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 758.616343] env[62460]: DEBUG oslo_concurrency.lockutils [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "refresh_cache-ec2297bb-1d29-4ea9-90f7-f4f39716c103" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 758.616492] env[62460]: DEBUG oslo_concurrency.lockutils [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquired lock "refresh_cache-ec2297bb-1d29-4ea9-90f7-f4f39716c103" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.616652] env[62460]: DEBUG nova.network.neutron [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 758.655211] env[62460]: DEBUG nova.network.neutron [-] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.712893] env[62460]: DEBUG nova.compute.manager [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 758.737437] env[62460]: DEBUG nova.virt.hardware [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 758.737687] env[62460]: DEBUG nova.virt.hardware [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 758.737841] env[62460]: DEBUG nova.virt.hardware [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 758.738026] env[62460]: DEBUG nova.virt.hardware [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 758.738182] env[62460]: DEBUG nova.virt.hardware [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 758.738330] env[62460]: DEBUG nova.virt.hardware [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 758.738538] env[62460]: DEBUG nova.virt.hardware [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 758.738698] env[62460]: DEBUG nova.virt.hardware [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 758.738861] env[62460]: DEBUG nova.virt.hardware [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 758.739032] env[62460]: DEBUG nova.virt.hardware [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 758.739210] env[62460]: DEBUG nova.virt.hardware [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 758.739503] env[62460]: DEBUG nova.network.neutron [-] [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.741057] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0c8a638-5606-4dcb-ab31-d1763efcfeb1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.748864] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f120077-38eb-4c18-a4e6-a8d1294f7524 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.762914] env[62460]: ERROR nova.compute.manager [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 44c1e29f-674f-41c8-9ce8-737c88692e0b, please check neutron logs for more information. [ 758.762914] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] Traceback (most recent call last): [ 758.762914] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 758.762914] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] yield resources [ 758.762914] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 758.762914] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] self.driver.spawn(context, instance, image_meta, [ 758.762914] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 758.762914] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] self._vmops.spawn(context, instance, image_meta, injected_files, [ 758.762914] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 758.762914] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] vm_ref = self.build_virtual_machine(instance, [ 758.762914] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 758.763301] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] vif_infos = vmwarevif.get_vif_info(self._session, [ 758.763301] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 758.763301] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] for vif in network_info: [ 758.763301] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 758.763301] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] return self._sync_wrapper(fn, *args, **kwargs) [ 758.763301] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 758.763301] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] self.wait() [ 758.763301] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 758.763301] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] self[:] = self._gt.wait() [ 758.763301] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 758.763301] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] return self._exit_event.wait() [ 758.763301] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 758.763301] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] current.throw(*self._exc) [ 758.763736] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 758.763736] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] result = function(*args, **kwargs) [ 758.763736] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 758.763736] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] return func(*args, **kwargs) [ 758.763736] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 758.763736] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] raise e [ 758.763736] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 758.763736] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] nwinfo = self.network_api.allocate_for_instance( [ 758.763736] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 758.763736] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] created_port_ids = self._update_ports_for_instance( [ 758.763736] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 758.763736] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] with excutils.save_and_reraise_exception(): [ 758.763736] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 758.764195] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] self.force_reraise() [ 758.764195] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 758.764195] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] raise self.value [ 758.764195] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 758.764195] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] updated_port = self._update_port( [ 758.764195] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 758.764195] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] _ensure_no_port_binding_failure(port) [ 758.764195] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 758.764195] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] raise exception.PortBindingFailed(port_id=port['id']) [ 758.764195] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] nova.exception.PortBindingFailed: Binding failed for port 44c1e29f-674f-41c8-9ce8-737c88692e0b, please check neutron logs for more information. [ 758.764195] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] [ 758.764195] env[62460]: INFO nova.compute.manager [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] Terminating instance [ 758.765017] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] Acquiring lock "refresh_cache-4f3c7bd0-314a-45d2-a75d-a22b29d67cfe" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 758.765188] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] Acquired lock "refresh_cache-4f3c7bd0-314a-45d2-a75d-a22b29d67cfe" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.765355] env[62460]: DEBUG nova.network.neutron [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 759.146024] env[62460]: DEBUG nova.network.neutron [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 759.153383] env[62460]: DEBUG nova.network.neutron [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 4ed90d16-81a6-4dbd-8936-0e137151171f] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.158261] env[62460]: INFO nova.compute.manager [-] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Took 1.02 seconds to deallocate network for instance. [ 759.241008] env[62460]: DEBUG nova.network.neutron [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.245210] env[62460]: INFO nova.compute.manager [-] [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] Took 1.03 seconds to deallocate network for instance. [ 759.250027] env[62460]: DEBUG nova.compute.claims [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 759.250027] env[62460]: DEBUG oslo_concurrency.lockutils [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 759.285470] env[62460]: DEBUG nova.network.neutron [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 759.365118] env[62460]: DEBUG nova.network.neutron [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.446889] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19fbda28-8926-4c71-a873-87cb1bd50c85 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.454337] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d6f0332-4dd3-4e49-8933-f352a014bacd {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.483507] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f29bc7c8-ebc2-4c0d-a06a-547b35ff8786 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.490298] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88674edf-dae4-406f-95a6-22a2a2f7ad24 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.502890] env[62460]: DEBUG nova.compute.provider_tree [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 759.661401] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Releasing lock "refresh_cache-4ed90d16-81a6-4dbd-8936-0e137151171f" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 759.661619] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 4ed90d16-81a6-4dbd-8936-0e137151171f] Updated the network info_cache for instance {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 759.661815] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 759.661981] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 759.662148] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 759.662293] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 759.662444] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 759.662569] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62460) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 759.662709] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 759.667898] env[62460]: DEBUG oslo_concurrency.lockutils [None req-59e9b2ee-6b25-41d0-8af5-fad3a3964ef2 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 759.743999] env[62460]: DEBUG oslo_concurrency.lockutils [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Releasing lock "refresh_cache-ec2297bb-1d29-4ea9-90f7-f4f39716c103" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 759.743999] env[62460]: DEBUG nova.compute.manager [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 759.743999] env[62460]: DEBUG nova.compute.manager [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 759.744200] env[62460]: DEBUG nova.network.neutron [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 759.758629] env[62460]: DEBUG nova.network.neutron [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 759.868201] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] Releasing lock "refresh_cache-4f3c7bd0-314a-45d2-a75d-a22b29d67cfe" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 759.869464] env[62460]: DEBUG nova.compute.manager [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 759.869705] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 759.870444] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-def2f3b2-ad9c-4997-9eba-9f70bd9da41d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.880014] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a72405f-3bb6-44f0-b2ca-350e376f0452 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.902904] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe could not be found. [ 759.903166] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 759.903355] env[62460]: INFO nova.compute.manager [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] Took 0.03 seconds to destroy the instance on the hypervisor. [ 759.903634] env[62460]: DEBUG oslo.service.loopingcall [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 759.903860] env[62460]: DEBUG nova.compute.manager [-] [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 759.903970] env[62460]: DEBUG nova.network.neutron [-] [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 759.937643] env[62460]: DEBUG nova.network.neutron [-] [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 759.986413] env[62460]: DEBUG nova.compute.manager [req-58143ba0-021d-428d-9763-e0ea6691414b req-2d7a02c5-0ce8-4a2f-9668-8072c888411b service nova] [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] Received event network-changed-44c1e29f-674f-41c8-9ce8-737c88692e0b {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 759.986584] env[62460]: DEBUG nova.compute.manager [req-58143ba0-021d-428d-9763-e0ea6691414b req-2d7a02c5-0ce8-4a2f-9668-8072c888411b service nova] [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] Refreshing instance network info cache due to event network-changed-44c1e29f-674f-41c8-9ce8-737c88692e0b. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 759.986792] env[62460]: DEBUG oslo_concurrency.lockutils [req-58143ba0-021d-428d-9763-e0ea6691414b req-2d7a02c5-0ce8-4a2f-9668-8072c888411b service nova] Acquiring lock "refresh_cache-4f3c7bd0-314a-45d2-a75d-a22b29d67cfe" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 759.986933] env[62460]: DEBUG oslo_concurrency.lockutils [req-58143ba0-021d-428d-9763-e0ea6691414b req-2d7a02c5-0ce8-4a2f-9668-8072c888411b service nova] Acquired lock "refresh_cache-4f3c7bd0-314a-45d2-a75d-a22b29d67cfe" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.987135] env[62460]: DEBUG nova.network.neutron [req-58143ba0-021d-428d-9763-e0ea6691414b req-2d7a02c5-0ce8-4a2f-9668-8072c888411b service nova] [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] Refreshing network info cache for port 44c1e29f-674f-41c8-9ce8-737c88692e0b {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 760.006353] env[62460]: DEBUG nova.scheduler.client.report [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 760.165231] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.262259] env[62460]: DEBUG nova.network.neutron [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.440384] env[62460]: DEBUG nova.network.neutron [-] [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.510118] env[62460]: DEBUG nova.network.neutron [req-58143ba0-021d-428d-9763-e0ea6691414b req-2d7a02c5-0ce8-4a2f-9668-8072c888411b service nova] [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 760.512387] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.900s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 760.512968] env[62460]: ERROR nova.compute.manager [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ed57313b-d260-4e7a-99b1-e4a6318a31b5, please check neutron logs for more information. [ 760.512968] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] Traceback (most recent call last): [ 760.512968] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 760.512968] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] self.driver.spawn(context, instance, image_meta, [ 760.512968] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 760.512968] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] self._vmops.spawn(context, instance, image_meta, injected_files, [ 760.512968] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 760.512968] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] vm_ref = self.build_virtual_machine(instance, [ 760.512968] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 760.512968] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] vif_infos = vmwarevif.get_vif_info(self._session, [ 760.512968] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 760.513572] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] for vif in network_info: [ 760.513572] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 760.513572] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] return self._sync_wrapper(fn, *args, **kwargs) [ 760.513572] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 760.513572] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] self.wait() [ 760.513572] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 760.513572] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] self[:] = self._gt.wait() [ 760.513572] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 760.513572] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] return self._exit_event.wait() [ 760.513572] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 760.513572] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] result = hub.switch() [ 760.513572] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 760.513572] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] return self.greenlet.switch() [ 760.514224] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 760.514224] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] result = function(*args, **kwargs) [ 760.514224] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 760.514224] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] return func(*args, **kwargs) [ 760.514224] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 760.514224] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] raise e [ 760.514224] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 760.514224] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] nwinfo = self.network_api.allocate_for_instance( [ 760.514224] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 760.514224] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] created_port_ids = self._update_ports_for_instance( [ 760.514224] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 760.514224] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] with excutils.save_and_reraise_exception(): [ 760.514224] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 760.514702] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] self.force_reraise() [ 760.514702] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 760.514702] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] raise self.value [ 760.514702] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 760.514702] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] updated_port = self._update_port( [ 760.514702] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 760.514702] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] _ensure_no_port_binding_failure(port) [ 760.514702] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 760.514702] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] raise exception.PortBindingFailed(port_id=port['id']) [ 760.514702] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] nova.exception.PortBindingFailed: Binding failed for port ed57313b-d260-4e7a-99b1-e4a6318a31b5, please check neutron logs for more information. [ 760.514702] env[62460]: ERROR nova.compute.manager [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] [ 760.515108] env[62460]: DEBUG nova.compute.utils [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] Binding failed for port ed57313b-d260-4e7a-99b1-e4a6318a31b5, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 760.516033] env[62460]: DEBUG oslo_concurrency.lockutils [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.174s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 760.516761] env[62460]: INFO nova.compute.claims [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 760.519623] env[62460]: DEBUG nova.compute.manager [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] Build of instance 0e00ca48-f9e1-4999-aad5-d7965f7ddc28 was re-scheduled: Binding failed for port ed57313b-d260-4e7a-99b1-e4a6318a31b5, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 760.520074] env[62460]: DEBUG nova.compute.manager [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 760.520304] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] Acquiring lock "refresh_cache-0e00ca48-f9e1-4999-aad5-d7965f7ddc28" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 760.520453] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] Acquired lock "refresh_cache-0e00ca48-f9e1-4999-aad5-d7965f7ddc28" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 760.520615] env[62460]: DEBUG nova.network.neutron [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 760.589801] env[62460]: DEBUG nova.network.neutron [req-58143ba0-021d-428d-9763-e0ea6691414b req-2d7a02c5-0ce8-4a2f-9668-8072c888411b service nova] [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.766719] env[62460]: INFO nova.compute.manager [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: ec2297bb-1d29-4ea9-90f7-f4f39716c103] Took 1.02 seconds to deallocate network for instance. [ 760.946781] env[62460]: INFO nova.compute.manager [-] [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] Took 1.04 seconds to deallocate network for instance. [ 760.949033] env[62460]: DEBUG nova.compute.claims [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 760.949216] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 761.039466] env[62460]: DEBUG nova.network.neutron [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 761.092429] env[62460]: DEBUG oslo_concurrency.lockutils [req-58143ba0-021d-428d-9763-e0ea6691414b req-2d7a02c5-0ce8-4a2f-9668-8072c888411b service nova] Releasing lock "refresh_cache-4f3c7bd0-314a-45d2-a75d-a22b29d67cfe" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 761.093126] env[62460]: DEBUG nova.compute.manager [req-58143ba0-021d-428d-9763-e0ea6691414b req-2d7a02c5-0ce8-4a2f-9668-8072c888411b service nova] [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] Received event network-vif-deleted-44c1e29f-674f-41c8-9ce8-737c88692e0b {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 761.123751] env[62460]: DEBUG nova.network.neutron [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.625813] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] Releasing lock "refresh_cache-0e00ca48-f9e1-4999-aad5-d7965f7ddc28" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 761.626232] env[62460]: DEBUG nova.compute.manager [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 761.626330] env[62460]: DEBUG nova.compute.manager [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 761.626450] env[62460]: DEBUG nova.network.neutron [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 761.641902] env[62460]: DEBUG nova.network.neutron [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 761.791781] env[62460]: INFO nova.scheduler.client.report [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Deleted allocations for instance ec2297bb-1d29-4ea9-90f7-f4f39716c103 [ 761.811782] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b195f80-cd60-4279-94ed-6841e44ba3b7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.820554] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68f33752-a26f-4bbc-83c9-ad714a9a2064 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.851945] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-548eb661-8652-4622-8c8a-ec46164863d1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.859631] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3ea2786-71c9-4076-87c1-97b712ac439a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.873542] env[62460]: DEBUG nova.compute.provider_tree [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 762.144668] env[62460]: DEBUG nova.network.neutron [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.302560] env[62460]: DEBUG oslo_concurrency.lockutils [None req-dcaea1ec-a8d3-4ccf-8011-6b82acaf46f2 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "ec2297bb-1d29-4ea9-90f7-f4f39716c103" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 151.124s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.378878] env[62460]: DEBUG nova.scheduler.client.report [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 762.647925] env[62460]: INFO nova.compute.manager [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] [instance: 0e00ca48-f9e1-4999-aad5-d7965f7ddc28] Took 1.02 seconds to deallocate network for instance. [ 762.804103] env[62460]: DEBUG nova.compute.manager [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 762.881774] env[62460]: DEBUG oslo_concurrency.lockutils [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.366s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.882419] env[62460]: DEBUG nova.compute.manager [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 762.884939] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.187s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 763.327102] env[62460]: DEBUG oslo_concurrency.lockutils [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 763.392673] env[62460]: DEBUG nova.compute.utils [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 763.394429] env[62460]: DEBUG nova.compute.manager [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 763.394429] env[62460]: DEBUG nova.network.neutron [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 763.455667] env[62460]: DEBUG nova.policy [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fd48df0f15684e57a2454cf0d785e173', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7e3c7f5a39534b2790895e5e7933069a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 763.678022] env[62460]: INFO nova.scheduler.client.report [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] Deleted allocations for instance 0e00ca48-f9e1-4999-aad5-d7965f7ddc28 [ 763.684999] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d59a84af-f7fa-4bac-8f9a-222e3e8cf6fd {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.694089] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-843a5bf6-2693-4b34-9125-96329d542349 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.725748] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57b52f5c-8607-4923-b5e5-daf0ca2e8701 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.733998] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38aadb20-d39b-42ed-a354-74adc6ff0a08 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.748343] env[62460]: DEBUG nova.compute.provider_tree [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 763.868388] env[62460]: DEBUG nova.network.neutron [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] Successfully created port: 0fcaca39-5cf6-4276-9963-5a22a97e131f {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 763.899631] env[62460]: DEBUG nova.compute.manager [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 764.190797] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a34b22e7-cb1b-4485-98c3-94f38d454f1c tempest-ServerRescueTestJSONUnderV235-571961654 tempest-ServerRescueTestJSONUnderV235-571961654-project-member] Lock "0e00ca48-f9e1-4999-aad5-d7965f7ddc28" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 152.631s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.253023] env[62460]: DEBUG nova.scheduler.client.report [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 764.540871] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "d7d180f7-e1a8-46c8-ba8e-ca50dac474cc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 764.541133] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "d7d180f7-e1a8-46c8-ba8e-ca50dac474cc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 764.688190] env[62460]: DEBUG nova.compute.manager [req-acdc189c-a4cf-4406-beaf-1b4fa6f755e5 req-192e28f4-dc5d-4768-a191-0cc02a5c543c service nova] [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] Received event network-changed-0fcaca39-5cf6-4276-9963-5a22a97e131f {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 764.688437] env[62460]: DEBUG nova.compute.manager [req-acdc189c-a4cf-4406-beaf-1b4fa6f755e5 req-192e28f4-dc5d-4768-a191-0cc02a5c543c service nova] [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] Refreshing instance network info cache due to event network-changed-0fcaca39-5cf6-4276-9963-5a22a97e131f. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 764.688606] env[62460]: DEBUG oslo_concurrency.lockutils [req-acdc189c-a4cf-4406-beaf-1b4fa6f755e5 req-192e28f4-dc5d-4768-a191-0cc02a5c543c service nova] Acquiring lock "refresh_cache-559757ef-ee92-4b88-8631-6f743fb88bc6" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 764.688795] env[62460]: DEBUG oslo_concurrency.lockutils [req-acdc189c-a4cf-4406-beaf-1b4fa6f755e5 req-192e28f4-dc5d-4768-a191-0cc02a5c543c service nova] Acquired lock "refresh_cache-559757ef-ee92-4b88-8631-6f743fb88bc6" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 764.688963] env[62460]: DEBUG nova.network.neutron [req-acdc189c-a4cf-4406-beaf-1b4fa6f755e5 req-192e28f4-dc5d-4768-a191-0cc02a5c543c service nova] [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] Refreshing network info cache for port 0fcaca39-5cf6-4276-9963-5a22a97e131f {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 764.691809] env[62460]: DEBUG nova.compute.manager [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 764.755657] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.871s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.756357] env[62460]: ERROR nova.compute.manager [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 319af934-2bcb-4020-bc62-7039d1fc7eb4, please check neutron logs for more information. [ 764.756357] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] Traceback (most recent call last): [ 764.756357] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 764.756357] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] self.driver.spawn(context, instance, image_meta, [ 764.756357] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 764.756357] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] self._vmops.spawn(context, instance, image_meta, injected_files, [ 764.756357] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 764.756357] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] vm_ref = self.build_virtual_machine(instance, [ 764.756357] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 764.756357] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] vif_infos = vmwarevif.get_vif_info(self._session, [ 764.756357] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 764.756762] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] for vif in network_info: [ 764.756762] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 764.756762] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] return self._sync_wrapper(fn, *args, **kwargs) [ 764.756762] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 764.756762] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] self.wait() [ 764.756762] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 764.756762] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] self[:] = self._gt.wait() [ 764.756762] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 764.756762] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] return self._exit_event.wait() [ 764.756762] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 764.756762] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] result = hub.switch() [ 764.756762] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 764.756762] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] return self.greenlet.switch() [ 764.757231] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 764.757231] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] result = function(*args, **kwargs) [ 764.757231] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 764.757231] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] return func(*args, **kwargs) [ 764.757231] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 764.757231] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] raise e [ 764.757231] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 764.757231] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] nwinfo = self.network_api.allocate_for_instance( [ 764.757231] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 764.757231] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] created_port_ids = self._update_ports_for_instance( [ 764.757231] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 764.757231] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] with excutils.save_and_reraise_exception(): [ 764.757231] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 764.757761] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] self.force_reraise() [ 764.757761] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 764.757761] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] raise self.value [ 764.757761] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 764.757761] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] updated_port = self._update_port( [ 764.757761] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 764.757761] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] _ensure_no_port_binding_failure(port) [ 764.757761] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 764.757761] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] raise exception.PortBindingFailed(port_id=port['id']) [ 764.757761] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] nova.exception.PortBindingFailed: Binding failed for port 319af934-2bcb-4020-bc62-7039d1fc7eb4, please check neutron logs for more information. [ 764.757761] env[62460]: ERROR nova.compute.manager [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] [ 764.758162] env[62460]: DEBUG nova.compute.utils [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] Binding failed for port 319af934-2bcb-4020-bc62-7039d1fc7eb4, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 764.758611] env[62460]: DEBUG oslo_concurrency.lockutils [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.963s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 764.762420] env[62460]: DEBUG nova.compute.manager [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] Build of instance ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee was re-scheduled: Binding failed for port 319af934-2bcb-4020-bc62-7039d1fc7eb4, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 764.762858] env[62460]: DEBUG nova.compute.manager [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 764.763109] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] Acquiring lock "refresh_cache-ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 764.763490] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] Acquired lock "refresh_cache-ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 764.763490] env[62460]: DEBUG nova.network.neutron [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 764.910045] env[62460]: DEBUG nova.compute.manager [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 764.934590] env[62460]: ERROR nova.compute.manager [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0fcaca39-5cf6-4276-9963-5a22a97e131f, please check neutron logs for more information. [ 764.934590] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 764.934590] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 764.934590] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 764.934590] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 764.934590] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 764.934590] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 764.934590] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 764.934590] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 764.934590] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 764.934590] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 764.934590] env[62460]: ERROR nova.compute.manager raise self.value [ 764.934590] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 764.934590] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 764.934590] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 764.934590] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 764.935160] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 764.935160] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 764.935160] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0fcaca39-5cf6-4276-9963-5a22a97e131f, please check neutron logs for more information. [ 764.935160] env[62460]: ERROR nova.compute.manager [ 764.935160] env[62460]: Traceback (most recent call last): [ 764.935160] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 764.935160] env[62460]: listener.cb(fileno) [ 764.935160] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 764.935160] env[62460]: result = function(*args, **kwargs) [ 764.935160] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 764.935160] env[62460]: return func(*args, **kwargs) [ 764.935160] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 764.935160] env[62460]: raise e [ 764.935160] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 764.935160] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 764.935160] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 764.935160] env[62460]: created_port_ids = self._update_ports_for_instance( [ 764.935160] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 764.935160] env[62460]: with excutils.save_and_reraise_exception(): [ 764.935160] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 764.935160] env[62460]: self.force_reraise() [ 764.935160] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 764.935160] env[62460]: raise self.value [ 764.935160] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 764.935160] env[62460]: updated_port = self._update_port( [ 764.935160] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 764.935160] env[62460]: _ensure_no_port_binding_failure(port) [ 764.935160] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 764.935160] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 764.936175] env[62460]: nova.exception.PortBindingFailed: Binding failed for port 0fcaca39-5cf6-4276-9963-5a22a97e131f, please check neutron logs for more information. [ 764.936175] env[62460]: Removing descriptor: 17 [ 764.939996] env[62460]: DEBUG nova.virt.hardware [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 764.940288] env[62460]: DEBUG nova.virt.hardware [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 764.940488] env[62460]: DEBUG nova.virt.hardware [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 764.940712] env[62460]: DEBUG nova.virt.hardware [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 764.940891] env[62460]: DEBUG nova.virt.hardware [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 764.941079] env[62460]: DEBUG nova.virt.hardware [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 764.941330] env[62460]: DEBUG nova.virt.hardware [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 764.941535] env[62460]: DEBUG nova.virt.hardware [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 764.941748] env[62460]: DEBUG nova.virt.hardware [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 764.941952] env[62460]: DEBUG nova.virt.hardware [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 764.942176] env[62460]: DEBUG nova.virt.hardware [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 764.943083] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-467593d9-7b36-4fec-9062-0d5b8358786b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.951573] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2369028-48ab-4bb8-b26b-ca98f5a72c54 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.965133] env[62460]: ERROR nova.compute.manager [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0fcaca39-5cf6-4276-9963-5a22a97e131f, please check neutron logs for more information. [ 764.965133] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] Traceback (most recent call last): [ 764.965133] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 764.965133] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] yield resources [ 764.965133] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 764.965133] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] self.driver.spawn(context, instance, image_meta, [ 764.965133] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 764.965133] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 764.965133] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 764.965133] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] vm_ref = self.build_virtual_machine(instance, [ 764.965133] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 764.965573] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] vif_infos = vmwarevif.get_vif_info(self._session, [ 764.965573] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 764.965573] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] for vif in network_info: [ 764.965573] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 764.965573] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] return self._sync_wrapper(fn, *args, **kwargs) [ 764.965573] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 764.965573] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] self.wait() [ 764.965573] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 764.965573] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] self[:] = self._gt.wait() [ 764.965573] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 764.965573] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] return self._exit_event.wait() [ 764.965573] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 764.965573] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] current.throw(*self._exc) [ 764.965972] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 764.965972] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] result = function(*args, **kwargs) [ 764.965972] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 764.965972] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] return func(*args, **kwargs) [ 764.965972] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 764.965972] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] raise e [ 764.965972] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 764.965972] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] nwinfo = self.network_api.allocate_for_instance( [ 764.965972] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 764.965972] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] created_port_ids = self._update_ports_for_instance( [ 764.965972] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 764.965972] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] with excutils.save_and_reraise_exception(): [ 764.965972] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 764.966397] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] self.force_reraise() [ 764.966397] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 764.966397] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] raise self.value [ 764.966397] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 764.966397] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] updated_port = self._update_port( [ 764.966397] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 764.966397] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] _ensure_no_port_binding_failure(port) [ 764.966397] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 764.966397] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] raise exception.PortBindingFailed(port_id=port['id']) [ 764.966397] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] nova.exception.PortBindingFailed: Binding failed for port 0fcaca39-5cf6-4276-9963-5a22a97e131f, please check neutron logs for more information. [ 764.966397] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] [ 764.966397] env[62460]: INFO nova.compute.manager [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] Terminating instance [ 764.967415] env[62460]: DEBUG oslo_concurrency.lockutils [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] Acquiring lock "refresh_cache-559757ef-ee92-4b88-8631-6f743fb88bc6" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 765.209306] env[62460]: DEBUG nova.network.neutron [req-acdc189c-a4cf-4406-beaf-1b4fa6f755e5 req-192e28f4-dc5d-4768-a191-0cc02a5c543c service nova] [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 765.218835] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.282702] env[62460]: DEBUG nova.network.neutron [req-acdc189c-a4cf-4406-beaf-1b4fa6f755e5 req-192e28f4-dc5d-4768-a191-0cc02a5c543c service nova] [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 765.361029] env[62460]: DEBUG nova.network.neutron [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 765.511235] env[62460]: DEBUG nova.network.neutron [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 765.585892] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e43d28b-fe5a-4b1a-968b-fdaafa539068 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.593547] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-574e57eb-6549-4b96-9214-8a7b2ccef079 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.624550] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62fceb70-caf6-4d95-bc01-323c21fd57c5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.631714] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35119a9b-a4c6-4cde-90b9-69233b212185 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.644496] env[62460]: DEBUG nova.compute.provider_tree [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 765.785700] env[62460]: DEBUG oslo_concurrency.lockutils [req-acdc189c-a4cf-4406-beaf-1b4fa6f755e5 req-192e28f4-dc5d-4768-a191-0cc02a5c543c service nova] Releasing lock "refresh_cache-559757ef-ee92-4b88-8631-6f743fb88bc6" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 765.785973] env[62460]: DEBUG oslo_concurrency.lockutils [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] Acquired lock "refresh_cache-559757ef-ee92-4b88-8631-6f743fb88bc6" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.785973] env[62460]: DEBUG nova.network.neutron [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 766.014187] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] Releasing lock "refresh_cache-ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 766.014503] env[62460]: DEBUG nova.compute.manager [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 766.014736] env[62460]: DEBUG nova.compute.manager [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 766.014960] env[62460]: DEBUG nova.network.neutron [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 766.030627] env[62460]: DEBUG nova.network.neutron [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 766.150025] env[62460]: DEBUG nova.scheduler.client.report [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 766.305453] env[62460]: DEBUG nova.network.neutron [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 766.360947] env[62460]: DEBUG nova.network.neutron [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.533960] env[62460]: DEBUG nova.network.neutron [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.653217] env[62460]: DEBUG oslo_concurrency.lockutils [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.894s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 766.653935] env[62460]: ERROR nova.compute.manager [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6fc7ca35-5bb9-4d46-9356-ceb957531a76, please check neutron logs for more information. [ 766.653935] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] Traceback (most recent call last): [ 766.653935] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 766.653935] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] self.driver.spawn(context, instance, image_meta, [ 766.653935] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 766.653935] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 766.653935] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 766.653935] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] vm_ref = self.build_virtual_machine(instance, [ 766.653935] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 766.653935] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] vif_infos = vmwarevif.get_vif_info(self._session, [ 766.653935] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 766.654498] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] for vif in network_info: [ 766.654498] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 766.654498] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] return self._sync_wrapper(fn, *args, **kwargs) [ 766.654498] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 766.654498] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] self.wait() [ 766.654498] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 766.654498] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] self[:] = self._gt.wait() [ 766.654498] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 766.654498] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] return self._exit_event.wait() [ 766.654498] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 766.654498] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] current.throw(*self._exc) [ 766.654498] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 766.654498] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] result = function(*args, **kwargs) [ 766.655132] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 766.655132] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] return func(*args, **kwargs) [ 766.655132] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 766.655132] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] raise e [ 766.655132] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 766.655132] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] nwinfo = self.network_api.allocate_for_instance( [ 766.655132] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 766.655132] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] created_port_ids = self._update_ports_for_instance( [ 766.655132] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 766.655132] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] with excutils.save_and_reraise_exception(): [ 766.655132] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 766.655132] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] self.force_reraise() [ 766.655132] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 766.655601] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] raise self.value [ 766.655601] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 766.655601] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] updated_port = self._update_port( [ 766.655601] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 766.655601] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] _ensure_no_port_binding_failure(port) [ 766.655601] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 766.655601] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] raise exception.PortBindingFailed(port_id=port['id']) [ 766.655601] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] nova.exception.PortBindingFailed: Binding failed for port 6fc7ca35-5bb9-4d46-9356-ceb957531a76, please check neutron logs for more information. [ 766.655601] env[62460]: ERROR nova.compute.manager [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] [ 766.655601] env[62460]: DEBUG nova.compute.utils [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] Binding failed for port 6fc7ca35-5bb9-4d46-9356-ceb957531a76, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 766.656065] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.712s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 766.657550] env[62460]: INFO nova.compute.claims [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 766.660274] env[62460]: DEBUG nova.compute.manager [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] Build of instance c4c5abc4-a5ed-4337-a930-d37d579819f8 was re-scheduled: Binding failed for port 6fc7ca35-5bb9-4d46-9356-ceb957531a76, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 766.660717] env[62460]: DEBUG nova.compute.manager [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 766.660953] env[62460]: DEBUG oslo_concurrency.lockutils [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] Acquiring lock "refresh_cache-c4c5abc4-a5ed-4337-a930-d37d579819f8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 766.661170] env[62460]: DEBUG oslo_concurrency.lockutils [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] Acquired lock "refresh_cache-c4c5abc4-a5ed-4337-a930-d37d579819f8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.661278] env[62460]: DEBUG nova.network.neutron [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 766.715025] env[62460]: DEBUG nova.compute.manager [req-3c7308e8-5280-49a6-8ca7-fd58a8786e90 req-4d6a07c3-8100-461a-93d1-c55c4e6fd952 service nova] [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] Received event network-vif-deleted-0fcaca39-5cf6-4276-9963-5a22a97e131f {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 766.863923] env[62460]: DEBUG oslo_concurrency.lockutils [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] Releasing lock "refresh_cache-559757ef-ee92-4b88-8631-6f743fb88bc6" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 766.864370] env[62460]: DEBUG nova.compute.manager [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 766.864561] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 766.864893] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9155f924-1e82-45f8-a2d3-9f204e7b79a0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.874425] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e92f9f5-e85e-48e2-a9e0-7f6077765388 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.896350] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 559757ef-ee92-4b88-8631-6f743fb88bc6 could not be found. [ 766.896554] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 766.896740] env[62460]: INFO nova.compute.manager [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] Took 0.03 seconds to destroy the instance on the hypervisor. [ 766.896980] env[62460]: DEBUG oslo.service.loopingcall [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 766.897196] env[62460]: DEBUG nova.compute.manager [-] [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 766.897292] env[62460]: DEBUG nova.network.neutron [-] [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 766.911287] env[62460]: DEBUG nova.network.neutron [-] [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 767.036080] env[62460]: INFO nova.compute.manager [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] [instance: ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee] Took 1.02 seconds to deallocate network for instance. [ 767.179957] env[62460]: DEBUG nova.network.neutron [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 767.251061] env[62460]: DEBUG nova.network.neutron [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.413497] env[62460]: DEBUG nova.network.neutron [-] [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.755889] env[62460]: DEBUG oslo_concurrency.lockutils [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] Releasing lock "refresh_cache-c4c5abc4-a5ed-4337-a930-d37d579819f8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 767.756191] env[62460]: DEBUG nova.compute.manager [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 767.756381] env[62460]: DEBUG nova.compute.manager [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 767.756551] env[62460]: DEBUG nova.network.neutron [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 767.773827] env[62460]: DEBUG nova.network.neutron [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 767.916330] env[62460]: INFO nova.compute.manager [-] [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] Took 1.02 seconds to deallocate network for instance. [ 767.918913] env[62460]: DEBUG nova.compute.claims [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 767.919139] env[62460]: DEBUG oslo_concurrency.lockutils [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.939088] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ac2e4aa-8f8a-496d-8fb9-65285dc6cbdc {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.954235] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1c31a03-e94b-4bb9-9590-7b1df868d2aa {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.984751] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c9ae072-308b-4a22-bb1a-f22452e06750 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.991831] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62a172e8-4269-4444-8280-4fa246afd3dd {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.004729] env[62460]: DEBUG nova.compute.provider_tree [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 768.065121] env[62460]: INFO nova.scheduler.client.report [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] Deleted allocations for instance ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee [ 768.276701] env[62460]: DEBUG nova.network.neutron [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 768.508212] env[62460]: DEBUG nova.scheduler.client.report [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 768.575997] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e36c3dd8-215f-47e9-86da-ac5dc2be4360 tempest-FloatingIPsAssociationTestJSON-871600147 tempest-FloatingIPsAssociationTestJSON-871600147-project-member] Lock "ecaa6bf7-e7bd-44c2-811c-1edf9a1039ee" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 141.817s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.778931] env[62460]: INFO nova.compute.manager [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] [instance: c4c5abc4-a5ed-4337-a930-d37d579819f8] Took 1.02 seconds to deallocate network for instance. [ 769.012812] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.357s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.013942] env[62460]: DEBUG nova.compute.manager [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 769.019637] env[62460]: DEBUG oslo_concurrency.lockutils [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.253s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 769.081597] env[62460]: DEBUG nova.compute.manager [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 769.523438] env[62460]: DEBUG nova.compute.utils [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 769.525175] env[62460]: DEBUG nova.compute.manager [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 769.525356] env[62460]: DEBUG nova.network.neutron [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 769.612279] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 769.624866] env[62460]: DEBUG nova.policy [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0f21627a28854b368512373c5931e51f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b76c231415f742e0a7a05dd04eee5fbc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 769.813884] env[62460]: INFO nova.scheduler.client.report [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] Deleted allocations for instance c4c5abc4-a5ed-4337-a930-d37d579819f8 [ 769.930020] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0773ce7b-2868-45df-b24a-a33a88e68807 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.937938] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33c0d1b3-8f33-4ad9-a906-93ae2f5dc61f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.725117] env[62460]: DEBUG nova.compute.manager [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 770.726904] env[62460]: DEBUG oslo_concurrency.lockutils [None req-946e2991-b25f-4494-91e0-d90a24a0cd42 tempest-ServerMetadataTestJSON-788976679 tempest-ServerMetadataTestJSON-788976679-project-member] Lock "c4c5abc4-a5ed-4337-a930-d37d579819f8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 140.158s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 770.727973] env[62460]: DEBUG nova.network.neutron [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] Successfully created port: b4e718dc-828e-49ea-8bd6-83b0a34f6d57 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 770.731193] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99f6c834-8684-4207-b730-5bb3aa3ed3d0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.740136] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2115c683-a989-42c3-ae50-7f638ad27563 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.754466] env[62460]: DEBUG nova.compute.provider_tree [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 771.237855] env[62460]: DEBUG nova.compute.manager [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 771.258217] env[62460]: DEBUG nova.scheduler.client.report [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 771.738861] env[62460]: DEBUG nova.compute.manager [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 771.743184] env[62460]: DEBUG nova.compute.manager [req-5dffcb0d-d71d-4b79-9bd8-a160a03dedc5 req-834d3cec-32bc-4938-a8f9-585d0ebaa660 service nova] [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] Received event network-changed-b4e718dc-828e-49ea-8bd6-83b0a34f6d57 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 771.743313] env[62460]: DEBUG nova.compute.manager [req-5dffcb0d-d71d-4b79-9bd8-a160a03dedc5 req-834d3cec-32bc-4938-a8f9-585d0ebaa660 service nova] [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] Refreshing instance network info cache due to event network-changed-b4e718dc-828e-49ea-8bd6-83b0a34f6d57. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 771.743533] env[62460]: DEBUG oslo_concurrency.lockutils [req-5dffcb0d-d71d-4b79-9bd8-a160a03dedc5 req-834d3cec-32bc-4938-a8f9-585d0ebaa660 service nova] Acquiring lock "refresh_cache-588e5489-2f61-473a-90e9-883eb8b163c8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 771.743683] env[62460]: DEBUG oslo_concurrency.lockutils [req-5dffcb0d-d71d-4b79-9bd8-a160a03dedc5 req-834d3cec-32bc-4938-a8f9-585d0ebaa660 service nova] Acquired lock "refresh_cache-588e5489-2f61-473a-90e9-883eb8b163c8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.743849] env[62460]: DEBUG nova.network.neutron [req-5dffcb0d-d71d-4b79-9bd8-a160a03dedc5 req-834d3cec-32bc-4938-a8f9-585d0ebaa660 service nova] [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] Refreshing network info cache for port b4e718dc-828e-49ea-8bd6-83b0a34f6d57 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 771.763478] env[62460]: DEBUG oslo_concurrency.lockutils [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.764082] env[62460]: DEBUG oslo_concurrency.lockutils [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.748s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.764590] env[62460]: ERROR nova.compute.manager [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a9889184-6641-40e8-a98a-0942457eed29, please check neutron logs for more information. [ 771.764590] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] Traceback (most recent call last): [ 771.764590] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 771.764590] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] self.driver.spawn(context, instance, image_meta, [ 771.764590] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 771.764590] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] self._vmops.spawn(context, instance, image_meta, injected_files, [ 771.764590] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 771.764590] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] vm_ref = self.build_virtual_machine(instance, [ 771.764590] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 771.764590] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] vif_infos = vmwarevif.get_vif_info(self._session, [ 771.764590] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 771.765146] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] for vif in network_info: [ 771.765146] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 771.765146] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] return self._sync_wrapper(fn, *args, **kwargs) [ 771.765146] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 771.765146] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] self.wait() [ 771.765146] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 771.765146] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] self[:] = self._gt.wait() [ 771.765146] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 771.765146] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] return self._exit_event.wait() [ 771.765146] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 771.765146] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] current.throw(*self._exc) [ 771.765146] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 771.765146] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] result = function(*args, **kwargs) [ 771.765755] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 771.765755] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] return func(*args, **kwargs) [ 771.765755] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 771.765755] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] raise e [ 771.765755] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 771.765755] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] nwinfo = self.network_api.allocate_for_instance( [ 771.765755] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 771.765755] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] created_port_ids = self._update_ports_for_instance( [ 771.765755] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 771.765755] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] with excutils.save_and_reraise_exception(): [ 771.765755] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 771.765755] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] self.force_reraise() [ 771.765755] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 771.766663] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] raise self.value [ 771.766663] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 771.766663] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] updated_port = self._update_port( [ 771.766663] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 771.766663] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] _ensure_no_port_binding_failure(port) [ 771.766663] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 771.766663] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] raise exception.PortBindingFailed(port_id=port['id']) [ 771.766663] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] nova.exception.PortBindingFailed: Binding failed for port a9889184-6641-40e8-a98a-0942457eed29, please check neutron logs for more information. [ 771.766663] env[62460]: ERROR nova.compute.manager [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] [ 771.766663] env[62460]: DEBUG nova.compute.utils [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] Binding failed for port a9889184-6641-40e8-a98a-0942457eed29, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 771.768448] env[62460]: DEBUG nova.virt.hardware [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 771.768703] env[62460]: DEBUG nova.virt.hardware [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 771.768821] env[62460]: DEBUG nova.virt.hardware [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 771.769015] env[62460]: DEBUG nova.virt.hardware [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 771.769169] env[62460]: DEBUG nova.virt.hardware [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 771.769317] env[62460]: DEBUG nova.virt.hardware [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 771.769520] env[62460]: DEBUG nova.virt.hardware [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 771.769678] env[62460]: DEBUG nova.virt.hardware [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 771.769841] env[62460]: DEBUG nova.virt.hardware [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 771.770011] env[62460]: DEBUG nova.virt.hardware [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 771.770190] env[62460]: DEBUG nova.virt.hardware [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 771.770572] env[62460]: DEBUG oslo_concurrency.lockutils [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.472s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.771960] env[62460]: INFO nova.compute.claims [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 771.775055] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9a5d794-7136-43eb-acfc-43a41bf655eb {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.778336] env[62460]: DEBUG nova.compute.manager [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] Build of instance e472b96a-5d82-4fb9-8d28-0901b71dc783 was re-scheduled: Binding failed for port a9889184-6641-40e8-a98a-0942457eed29, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 771.778771] env[62460]: DEBUG nova.compute.manager [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 771.779102] env[62460]: DEBUG oslo_concurrency.lockutils [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "refresh_cache-e472b96a-5d82-4fb9-8d28-0901b71dc783" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 771.779192] env[62460]: DEBUG oslo_concurrency.lockutils [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquired lock "refresh_cache-e472b96a-5d82-4fb9-8d28-0901b71dc783" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.779300] env[62460]: DEBUG nova.network.neutron [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 771.785971] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60d1c254-bff3-4af7-b5cc-689f103ae077 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.917882] env[62460]: ERROR nova.compute.manager [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b4e718dc-828e-49ea-8bd6-83b0a34f6d57, please check neutron logs for more information. [ 771.917882] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 771.917882] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 771.917882] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 771.917882] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 771.917882] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 771.917882] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 771.917882] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 771.917882] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 771.917882] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 771.917882] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 771.917882] env[62460]: ERROR nova.compute.manager raise self.value [ 771.917882] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 771.917882] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 771.917882] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 771.917882] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 771.918400] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 771.918400] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 771.918400] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b4e718dc-828e-49ea-8bd6-83b0a34f6d57, please check neutron logs for more information. [ 771.918400] env[62460]: ERROR nova.compute.manager [ 771.918400] env[62460]: Traceback (most recent call last): [ 771.918400] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 771.918400] env[62460]: listener.cb(fileno) [ 771.918400] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 771.918400] env[62460]: result = function(*args, **kwargs) [ 771.918400] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 771.918400] env[62460]: return func(*args, **kwargs) [ 771.918400] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 771.918400] env[62460]: raise e [ 771.918400] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 771.918400] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 771.918400] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 771.918400] env[62460]: created_port_ids = self._update_ports_for_instance( [ 771.918400] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 771.918400] env[62460]: with excutils.save_and_reraise_exception(): [ 771.918400] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 771.918400] env[62460]: self.force_reraise() [ 771.918400] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 771.918400] env[62460]: raise self.value [ 771.918400] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 771.918400] env[62460]: updated_port = self._update_port( [ 771.918400] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 771.918400] env[62460]: _ensure_no_port_binding_failure(port) [ 771.918400] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 771.918400] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 771.919260] env[62460]: nova.exception.PortBindingFailed: Binding failed for port b4e718dc-828e-49ea-8bd6-83b0a34f6d57, please check neutron logs for more information. [ 771.919260] env[62460]: Removing descriptor: 18 [ 771.919260] env[62460]: ERROR nova.compute.manager [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b4e718dc-828e-49ea-8bd6-83b0a34f6d57, please check neutron logs for more information. [ 771.919260] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] Traceback (most recent call last): [ 771.919260] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 771.919260] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] yield resources [ 771.919260] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 771.919260] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] self.driver.spawn(context, instance, image_meta, [ 771.919260] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 771.919260] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 771.919260] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 771.919260] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] vm_ref = self.build_virtual_machine(instance, [ 771.919683] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 771.919683] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] vif_infos = vmwarevif.get_vif_info(self._session, [ 771.919683] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 771.919683] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] for vif in network_info: [ 771.919683] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 771.919683] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] return self._sync_wrapper(fn, *args, **kwargs) [ 771.919683] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 771.919683] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] self.wait() [ 771.919683] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 771.919683] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] self[:] = self._gt.wait() [ 771.919683] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 771.919683] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] return self._exit_event.wait() [ 771.919683] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 771.920088] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] result = hub.switch() [ 771.920088] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 771.920088] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] return self.greenlet.switch() [ 771.920088] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 771.920088] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] result = function(*args, **kwargs) [ 771.920088] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 771.920088] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] return func(*args, **kwargs) [ 771.920088] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 771.920088] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] raise e [ 771.920088] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 771.920088] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] nwinfo = self.network_api.allocate_for_instance( [ 771.920088] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 771.920088] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] created_port_ids = self._update_ports_for_instance( [ 771.920472] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 771.920472] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] with excutils.save_and_reraise_exception(): [ 771.920472] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 771.920472] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] self.force_reraise() [ 771.920472] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 771.920472] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] raise self.value [ 771.920472] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 771.920472] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] updated_port = self._update_port( [ 771.920472] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 771.920472] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] _ensure_no_port_binding_failure(port) [ 771.920472] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 771.920472] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] raise exception.PortBindingFailed(port_id=port['id']) [ 771.920845] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] nova.exception.PortBindingFailed: Binding failed for port b4e718dc-828e-49ea-8bd6-83b0a34f6d57, please check neutron logs for more information. [ 771.920845] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] [ 771.920845] env[62460]: INFO nova.compute.manager [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] Terminating instance [ 771.921364] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Acquiring lock "refresh_cache-588e5489-2f61-473a-90e9-883eb8b163c8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 772.309724] env[62460]: DEBUG nova.network.neutron [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 772.423934] env[62460]: DEBUG nova.network.neutron [req-5dffcb0d-d71d-4b79-9bd8-a160a03dedc5 req-834d3cec-32bc-4938-a8f9-585d0ebaa660 service nova] [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 772.454938] env[62460]: DEBUG nova.network.neutron [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 772.558199] env[62460]: DEBUG nova.network.neutron [req-5dffcb0d-d71d-4b79-9bd8-a160a03dedc5 req-834d3cec-32bc-4938-a8f9-585d0ebaa660 service nova] [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 772.960712] env[62460]: DEBUG oslo_concurrency.lockutils [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Releasing lock "refresh_cache-e472b96a-5d82-4fb9-8d28-0901b71dc783" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 772.960990] env[62460]: DEBUG nova.compute.manager [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 772.961045] env[62460]: DEBUG nova.compute.manager [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 772.961289] env[62460]: DEBUG nova.network.neutron [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 772.984695] env[62460]: DEBUG nova.network.neutron [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 773.062115] env[62460]: DEBUG oslo_concurrency.lockutils [req-5dffcb0d-d71d-4b79-9bd8-a160a03dedc5 req-834d3cec-32bc-4938-a8f9-585d0ebaa660 service nova] Releasing lock "refresh_cache-588e5489-2f61-473a-90e9-883eb8b163c8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 773.063596] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Acquired lock "refresh_cache-588e5489-2f61-473a-90e9-883eb8b163c8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.063822] env[62460]: DEBUG nova.network.neutron [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 773.150295] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02a05662-5d6b-4c35-9c8f-00772d0f2aeb {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.158619] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e7d85f6-0acc-4e8d-8532-10b06998b998 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.195913] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cba0a8e-9ee7-4d20-9a35-6be9407f26f4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.203722] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81aec95d-93de-4001-8d14-4c0d3da36ef1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.220293] env[62460]: DEBUG nova.compute.provider_tree [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 773.495408] env[62460]: DEBUG nova.network.neutron [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.585044] env[62460]: DEBUG nova.network.neutron [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 773.673794] env[62460]: DEBUG nova.network.neutron [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.727036] env[62460]: DEBUG nova.scheduler.client.report [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 773.790093] env[62460]: DEBUG nova.compute.manager [req-44c448f4-850a-4eed-a065-009fe9247853 req-ece4f3d7-8f89-46d7-b321-28be8cfaa403 service nova] [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] Received event network-vif-deleted-b4e718dc-828e-49ea-8bd6-83b0a34f6d57 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 773.997157] env[62460]: INFO nova.compute.manager [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e472b96a-5d82-4fb9-8d28-0901b71dc783] Took 1.04 seconds to deallocate network for instance. [ 774.177951] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Releasing lock "refresh_cache-588e5489-2f61-473a-90e9-883eb8b163c8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 774.178379] env[62460]: DEBUG nova.compute.manager [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 774.178691] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 774.179378] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-51a509e5-124a-4123-a2f9-cf9f6a345627 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.188841] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-238d4c3f-0fb4-468d-aa7f-cff9d2fbe802 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.213658] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 588e5489-2f61-473a-90e9-883eb8b163c8 could not be found. [ 774.213853] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 774.213929] env[62460]: INFO nova.compute.manager [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] Took 0.04 seconds to destroy the instance on the hypervisor. [ 774.214244] env[62460]: DEBUG oslo.service.loopingcall [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 774.215693] env[62460]: DEBUG nova.compute.manager [-] [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 774.215693] env[62460]: DEBUG nova.network.neutron [-] [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 774.234365] env[62460]: DEBUG oslo_concurrency.lockutils [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.464s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 774.234891] env[62460]: DEBUG nova.compute.manager [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 774.240063] env[62460]: DEBUG nova.network.neutron [-] [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 774.240614] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 18.903s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 774.240825] env[62460]: DEBUG nova.objects.instance [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62460) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 774.745618] env[62460]: DEBUG nova.compute.utils [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 774.750271] env[62460]: DEBUG nova.network.neutron [-] [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.750429] env[62460]: DEBUG nova.compute.manager [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 774.750569] env[62460]: DEBUG nova.network.neutron [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 774.835931] env[62460]: DEBUG nova.policy [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'caef87e6a8554d1ca7eb12e2db86e70e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'acf65d070f164356a7de6afc03ed2995', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 775.045983] env[62460]: INFO nova.scheduler.client.report [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Deleted allocations for instance e472b96a-5d82-4fb9-8d28-0901b71dc783 [ 775.252205] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d86f0387-34d2-4906-b0f1-004f9c123f93 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.011s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.254227] env[62460]: DEBUG nova.compute.manager [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 775.257971] env[62460]: DEBUG oslo_concurrency.lockutils [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.009s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 775.260981] env[62460]: INFO nova.compute.manager [-] [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] Took 1.05 seconds to deallocate network for instance. [ 775.263429] env[62460]: DEBUG nova.compute.claims [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 775.263531] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 775.362354] env[62460]: DEBUG nova.network.neutron [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] Successfully created port: d957de37-a110-4cb6-b807-78c130f4de9e {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 775.547893] env[62460]: DEBUG oslo_concurrency.lockutils [None req-14dd2a0b-fc27-4f52-8d06-a99329a40c35 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "e472b96a-5d82-4fb9-8d28-0901b71dc783" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 111.230s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.055375] env[62460]: DEBUG nova.compute.manager [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 776.150713] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2639137e-7029-4c82-b881-16696dc127eb {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.158551] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-474bd299-eb7d-4b0f-8d8c-34e2b695294c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.190534] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b50698f0-a472-4a31-9141-5f50ad79fabd {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.198698] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79865fc3-4344-4cf5-aca6-e02eadc6de93 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.215809] env[62460]: DEBUG nova.compute.provider_tree [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 776.270132] env[62460]: DEBUG nova.compute.manager [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 776.301056] env[62460]: DEBUG nova.virt.hardware [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 776.301318] env[62460]: DEBUG nova.virt.hardware [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 776.301476] env[62460]: DEBUG nova.virt.hardware [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 776.301657] env[62460]: DEBUG nova.virt.hardware [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 776.301848] env[62460]: DEBUG nova.virt.hardware [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 776.301989] env[62460]: DEBUG nova.virt.hardware [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 776.302233] env[62460]: DEBUG nova.virt.hardware [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 776.302397] env[62460]: DEBUG nova.virt.hardware [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 776.302573] env[62460]: DEBUG nova.virt.hardware [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 776.302734] env[62460]: DEBUG nova.virt.hardware [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 776.302901] env[62460]: DEBUG nova.virt.hardware [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 776.303770] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03800e96-69ca-458a-b6a3-590f50b18b9b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.311504] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21b934ba-7520-485f-a937-f9d42af55566 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.382260] env[62460]: DEBUG nova.compute.manager [req-f0e19ce3-a50f-4779-ba15-50b32168b18b req-2c67f270-4262-4006-bb79-32439c300fae service nova] [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] Received event network-changed-d957de37-a110-4cb6-b807-78c130f4de9e {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 776.382463] env[62460]: DEBUG nova.compute.manager [req-f0e19ce3-a50f-4779-ba15-50b32168b18b req-2c67f270-4262-4006-bb79-32439c300fae service nova] [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] Refreshing instance network info cache due to event network-changed-d957de37-a110-4cb6-b807-78c130f4de9e. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 776.382677] env[62460]: DEBUG oslo_concurrency.lockutils [req-f0e19ce3-a50f-4779-ba15-50b32168b18b req-2c67f270-4262-4006-bb79-32439c300fae service nova] Acquiring lock "refresh_cache-6c58f9fe-fb6d-4012-8b69-39e4134996b2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.382820] env[62460]: DEBUG oslo_concurrency.lockutils [req-f0e19ce3-a50f-4779-ba15-50b32168b18b req-2c67f270-4262-4006-bb79-32439c300fae service nova] Acquired lock "refresh_cache-6c58f9fe-fb6d-4012-8b69-39e4134996b2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.383008] env[62460]: DEBUG nova.network.neutron [req-f0e19ce3-a50f-4779-ba15-50b32168b18b req-2c67f270-4262-4006-bb79-32439c300fae service nova] [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] Refreshing network info cache for port d957de37-a110-4cb6-b807-78c130f4de9e {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 776.569896] env[62460]: ERROR nova.compute.manager [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d957de37-a110-4cb6-b807-78c130f4de9e, please check neutron logs for more information. [ 776.569896] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 776.569896] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 776.569896] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 776.569896] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 776.569896] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 776.569896] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 776.569896] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 776.569896] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 776.569896] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 776.569896] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 776.569896] env[62460]: ERROR nova.compute.manager raise self.value [ 776.569896] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 776.569896] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 776.569896] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 776.569896] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 776.570263] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 776.570263] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 776.570263] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d957de37-a110-4cb6-b807-78c130f4de9e, please check neutron logs for more information. [ 776.570263] env[62460]: ERROR nova.compute.manager [ 776.570263] env[62460]: Traceback (most recent call last): [ 776.570263] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 776.570263] env[62460]: listener.cb(fileno) [ 776.570263] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 776.570263] env[62460]: result = function(*args, **kwargs) [ 776.570263] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 776.570263] env[62460]: return func(*args, **kwargs) [ 776.570263] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 776.570263] env[62460]: raise e [ 776.570263] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 776.570263] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 776.570263] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 776.570263] env[62460]: created_port_ids = self._update_ports_for_instance( [ 776.570263] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 776.570263] env[62460]: with excutils.save_and_reraise_exception(): [ 776.570263] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 776.570263] env[62460]: self.force_reraise() [ 776.570263] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 776.570263] env[62460]: raise self.value [ 776.570263] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 776.570263] env[62460]: updated_port = self._update_port( [ 776.570263] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 776.570263] env[62460]: _ensure_no_port_binding_failure(port) [ 776.570263] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 776.570263] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 776.570873] env[62460]: nova.exception.PortBindingFailed: Binding failed for port d957de37-a110-4cb6-b807-78c130f4de9e, please check neutron logs for more information. [ 776.570873] env[62460]: Removing descriptor: 18 [ 776.571125] env[62460]: ERROR nova.compute.manager [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d957de37-a110-4cb6-b807-78c130f4de9e, please check neutron logs for more information. [ 776.571125] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] Traceback (most recent call last): [ 776.571125] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 776.571125] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] yield resources [ 776.571125] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 776.571125] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] self.driver.spawn(context, instance, image_meta, [ 776.571125] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 776.571125] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 776.571125] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 776.571125] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] vm_ref = self.build_virtual_machine(instance, [ 776.571125] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 776.571399] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] vif_infos = vmwarevif.get_vif_info(self._session, [ 776.571399] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 776.571399] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] for vif in network_info: [ 776.571399] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 776.571399] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] return self._sync_wrapper(fn, *args, **kwargs) [ 776.571399] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 776.571399] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] self.wait() [ 776.571399] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 776.571399] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] self[:] = self._gt.wait() [ 776.571399] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 776.571399] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] return self._exit_event.wait() [ 776.571399] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 776.571399] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] result = hub.switch() [ 776.571775] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 776.571775] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] return self.greenlet.switch() [ 776.571775] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 776.571775] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] result = function(*args, **kwargs) [ 776.571775] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 776.571775] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] return func(*args, **kwargs) [ 776.571775] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 776.571775] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] raise e [ 776.571775] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 776.571775] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] nwinfo = self.network_api.allocate_for_instance( [ 776.571775] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 776.571775] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] created_port_ids = self._update_ports_for_instance( [ 776.571775] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 776.572079] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] with excutils.save_and_reraise_exception(): [ 776.572079] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 776.572079] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] self.force_reraise() [ 776.572079] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 776.572079] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] raise self.value [ 776.572079] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 776.572079] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] updated_port = self._update_port( [ 776.572079] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 776.572079] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] _ensure_no_port_binding_failure(port) [ 776.572079] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 776.572079] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] raise exception.PortBindingFailed(port_id=port['id']) [ 776.572079] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] nova.exception.PortBindingFailed: Binding failed for port d957de37-a110-4cb6-b807-78c130f4de9e, please check neutron logs for more information. [ 776.572079] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] [ 776.572392] env[62460]: INFO nova.compute.manager [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] Terminating instance [ 776.573661] env[62460]: DEBUG oslo_concurrency.lockutils [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] Acquiring lock "refresh_cache-6c58f9fe-fb6d-4012-8b69-39e4134996b2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.577590] env[62460]: DEBUG oslo_concurrency.lockutils [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 776.648086] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "16d40829-ba6e-4193-98b5-fff94c066bcb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 776.648340] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "16d40829-ba6e-4193-98b5-fff94c066bcb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 776.721007] env[62460]: DEBUG nova.scheduler.client.report [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 776.900221] env[62460]: DEBUG nova.network.neutron [req-f0e19ce3-a50f-4779-ba15-50b32168b18b req-2c67f270-4262-4006-bb79-32439c300fae service nova] [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 776.982297] env[62460]: DEBUG nova.network.neutron [req-f0e19ce3-a50f-4779-ba15-50b32168b18b req-2c67f270-4262-4006-bb79-32439c300fae service nova] [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.225930] env[62460]: DEBUG oslo_concurrency.lockutils [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.969s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.226625] env[62460]: ERROR nova.compute.manager [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0df43069-277e-4946-8608-d75983f32608, please check neutron logs for more information. [ 777.226625] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] Traceback (most recent call last): [ 777.226625] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 777.226625] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] self.driver.spawn(context, instance, image_meta, [ 777.226625] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 777.226625] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] self._vmops.spawn(context, instance, image_meta, injected_files, [ 777.226625] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 777.226625] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] vm_ref = self.build_virtual_machine(instance, [ 777.226625] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 777.226625] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] vif_infos = vmwarevif.get_vif_info(self._session, [ 777.226625] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 777.226929] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] for vif in network_info: [ 777.226929] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 777.226929] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] return self._sync_wrapper(fn, *args, **kwargs) [ 777.226929] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 777.226929] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] self.wait() [ 777.226929] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 777.226929] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] self[:] = self._gt.wait() [ 777.226929] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 777.226929] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] return self._exit_event.wait() [ 777.226929] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 777.226929] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] current.throw(*self._exc) [ 777.226929] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 777.226929] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] result = function(*args, **kwargs) [ 777.227255] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 777.227255] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] return func(*args, **kwargs) [ 777.227255] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 777.227255] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] raise e [ 777.227255] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 777.227255] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] nwinfo = self.network_api.allocate_for_instance( [ 777.227255] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 777.227255] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] created_port_ids = self._update_ports_for_instance( [ 777.227255] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 777.227255] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] with excutils.save_and_reraise_exception(): [ 777.227255] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 777.227255] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] self.force_reraise() [ 777.227255] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 777.227576] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] raise self.value [ 777.227576] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 777.227576] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] updated_port = self._update_port( [ 777.227576] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 777.227576] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] _ensure_no_port_binding_failure(port) [ 777.227576] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 777.227576] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] raise exception.PortBindingFailed(port_id=port['id']) [ 777.227576] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] nova.exception.PortBindingFailed: Binding failed for port 0df43069-277e-4946-8608-d75983f32608, please check neutron logs for more information. [ 777.227576] env[62460]: ERROR nova.compute.manager [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] [ 777.227576] env[62460]: DEBUG nova.compute.utils [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] Binding failed for port 0df43069-277e-4946-8608-d75983f32608, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 777.228501] env[62460]: DEBUG oslo_concurrency.lockutils [None req-59e9b2ee-6b25-41d0-8af5-fad3a3964ef2 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.561s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.228719] env[62460]: DEBUG nova.objects.instance [None req-59e9b2ee-6b25-41d0-8af5-fad3a3964ef2 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Lazy-loading 'resources' on Instance uuid 5214f4a1-3e28-41bf-88d9-161511385e1b {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 777.229920] env[62460]: DEBUG nova.compute.manager [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] Build of instance be50b542-4c61-4da3-af4e-8a57e05a77ee was re-scheduled: Binding failed for port 0df43069-277e-4946-8608-d75983f32608, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 777.230322] env[62460]: DEBUG nova.compute.manager [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 777.230541] env[62460]: DEBUG oslo_concurrency.lockutils [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Acquiring lock "refresh_cache-be50b542-4c61-4da3-af4e-8a57e05a77ee" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.230685] env[62460]: DEBUG oslo_concurrency.lockutils [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Acquired lock "refresh_cache-be50b542-4c61-4da3-af4e-8a57e05a77ee" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.230844] env[62460]: DEBUG nova.network.neutron [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 777.484557] env[62460]: DEBUG oslo_concurrency.lockutils [req-f0e19ce3-a50f-4779-ba15-50b32168b18b req-2c67f270-4262-4006-bb79-32439c300fae service nova] Releasing lock "refresh_cache-6c58f9fe-fb6d-4012-8b69-39e4134996b2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 777.485014] env[62460]: DEBUG oslo_concurrency.lockutils [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] Acquired lock "refresh_cache-6c58f9fe-fb6d-4012-8b69-39e4134996b2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.485225] env[62460]: DEBUG nova.network.neutron [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 777.834421] env[62460]: DEBUG nova.network.neutron [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 777.894308] env[62460]: DEBUG nova.scheduler.client.report [None req-59e9b2ee-6b25-41d0-8af5-fad3a3964ef2 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Refreshing inventories for resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 777.909660] env[62460]: DEBUG nova.network.neutron [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.912581] env[62460]: DEBUG nova.scheduler.client.report [None req-59e9b2ee-6b25-41d0-8af5-fad3a3964ef2 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Updating ProviderTree inventory for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 777.912581] env[62460]: DEBUG nova.compute.provider_tree [None req-59e9b2ee-6b25-41d0-8af5-fad3a3964ef2 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 777.923220] env[62460]: DEBUG nova.scheduler.client.report [None req-59e9b2ee-6b25-41d0-8af5-fad3a3964ef2 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Refreshing aggregate associations for resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728, aggregates: None {{(pid=62460) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 777.941023] env[62460]: DEBUG nova.scheduler.client.report [None req-59e9b2ee-6b25-41d0-8af5-fad3a3964ef2 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Refreshing trait associations for resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62460) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 778.005224] env[62460]: DEBUG nova.network.neutron [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 778.095569] env[62460]: DEBUG nova.network.neutron [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.207256] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6355732-bed8-4991-ba6d-1e0c17b5b117 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.214934] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cc30660-45d7-46ff-9c7d-135c5f2e9455 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.244978] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ac70212-ca9d-4b88-bfc6-14c312cc0a2e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.252299] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5501e261-9e07-4d8a-b159-21b7b7d31f73 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.265251] env[62460]: DEBUG nova.compute.provider_tree [None req-59e9b2ee-6b25-41d0-8af5-fad3a3964ef2 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 778.416898] env[62460]: DEBUG oslo_concurrency.lockutils [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Releasing lock "refresh_cache-be50b542-4c61-4da3-af4e-8a57e05a77ee" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.417202] env[62460]: DEBUG nova.compute.manager [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 778.417430] env[62460]: DEBUG nova.compute.manager [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 778.417630] env[62460]: DEBUG nova.network.neutron [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 778.420065] env[62460]: DEBUG nova.compute.manager [req-d26832e1-f288-4a90-809e-a7e50b8b3077 req-71eec5b3-07a8-4833-bde2-d9ab56d73110 service nova] [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] Received event network-vif-deleted-d957de37-a110-4cb6-b807-78c130f4de9e {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 778.433453] env[62460]: DEBUG nova.network.neutron [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 778.599675] env[62460]: DEBUG oslo_concurrency.lockutils [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] Releasing lock "refresh_cache-6c58f9fe-fb6d-4012-8b69-39e4134996b2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.600116] env[62460]: DEBUG nova.compute.manager [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 778.600316] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 778.600611] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d228d055-3b45-4ce0-a2f1-059e0307bcd2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.609329] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dcd8f62-e5e1-489b-92cf-fe5116347f8d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.631143] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6c58f9fe-fb6d-4012-8b69-39e4134996b2 could not be found. [ 778.631374] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 778.631553] env[62460]: INFO nova.compute.manager [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] Took 0.03 seconds to destroy the instance on the hypervisor. [ 778.631791] env[62460]: DEBUG oslo.service.loopingcall [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 778.632031] env[62460]: DEBUG nova.compute.manager [-] [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 778.632131] env[62460]: DEBUG nova.network.neutron [-] [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 778.646174] env[62460]: DEBUG nova.network.neutron [-] [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 778.768685] env[62460]: DEBUG nova.scheduler.client.report [None req-59e9b2ee-6b25-41d0-8af5-fad3a3964ef2 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 778.935998] env[62460]: DEBUG nova.network.neutron [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.149488] env[62460]: DEBUG nova.network.neutron [-] [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.273108] env[62460]: DEBUG oslo_concurrency.lockutils [None req-59e9b2ee-6b25-41d0-8af5-fad3a3964ef2 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.044s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.275726] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 19.111s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.275903] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.276073] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62460) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 779.276356] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.327s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.279601] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afd57ba5-9fe5-4a37-9dc7-ffaa009e329e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.287562] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72e9c408-72d3-4a52-b04a-a1b4a31868c9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.292918] env[62460]: INFO nova.scheduler.client.report [None req-59e9b2ee-6b25-41d0-8af5-fad3a3964ef2 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Deleted allocations for instance 5214f4a1-3e28-41bf-88d9-161511385e1b [ 779.304743] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f09a389-a68e-4c40-8d75-100a484fb565 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.311823] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b911e76a-180d-4db3-bc05-3ed30b251173 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.343104] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181517MB free_disk=135GB free_vcpus=48 pci_devices=None {{(pid=62460) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 779.343104] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 779.438862] env[62460]: INFO nova.compute.manager [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: be50b542-4c61-4da3-af4e-8a57e05a77ee] Took 1.02 seconds to deallocate network for instance. [ 779.651882] env[62460]: INFO nova.compute.manager [-] [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] Took 1.02 seconds to deallocate network for instance. [ 779.654250] env[62460]: DEBUG nova.compute.claims [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 779.654434] env[62460]: DEBUG oslo_concurrency.lockutils [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 779.806871] env[62460]: DEBUG oslo_concurrency.lockutils [None req-59e9b2ee-6b25-41d0-8af5-fad3a3964ef2 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Lock "5214f4a1-3e28-41bf-88d9-161511385e1b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.841s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.012516] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89e04fcb-be8b-4cf7-bc6a-3d08229bd75d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.019847] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc8b29a0-c096-49ee-8d88-d94f842ea6c9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.048586] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f4f5bf5-c0cf-4063-a7aa-a8b9434a2d52 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.055302] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92ec4a52-27f8-4f0e-9537-198da1242445 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.067740] env[62460]: DEBUG nova.compute.provider_tree [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 780.464242] env[62460]: INFO nova.scheduler.client.report [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Deleted allocations for instance be50b542-4c61-4da3-af4e-8a57e05a77ee [ 780.570528] env[62460]: DEBUG nova.scheduler.client.report [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 780.799363] env[62460]: DEBUG oslo_concurrency.lockutils [None req-cb805937-3829-46aa-b450-89395080ba46 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Acquiring lock "4ed90d16-81a6-4dbd-8936-0e137151171f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 780.800022] env[62460]: DEBUG oslo_concurrency.lockutils [None req-cb805937-3829-46aa-b450-89395080ba46 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Lock "4ed90d16-81a6-4dbd-8936-0e137151171f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 780.800022] env[62460]: DEBUG oslo_concurrency.lockutils [None req-cb805937-3829-46aa-b450-89395080ba46 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Acquiring lock "4ed90d16-81a6-4dbd-8936-0e137151171f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 780.800022] env[62460]: DEBUG oslo_concurrency.lockutils [None req-cb805937-3829-46aa-b450-89395080ba46 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Lock "4ed90d16-81a6-4dbd-8936-0e137151171f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 780.800148] env[62460]: DEBUG oslo_concurrency.lockutils [None req-cb805937-3829-46aa-b450-89395080ba46 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Lock "4ed90d16-81a6-4dbd-8936-0e137151171f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.802305] env[62460]: INFO nova.compute.manager [None req-cb805937-3829-46aa-b450-89395080ba46 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 4ed90d16-81a6-4dbd-8936-0e137151171f] Terminating instance [ 780.804433] env[62460]: DEBUG oslo_concurrency.lockutils [None req-cb805937-3829-46aa-b450-89395080ba46 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Acquiring lock "refresh_cache-4ed90d16-81a6-4dbd-8936-0e137151171f" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 780.804598] env[62460]: DEBUG oslo_concurrency.lockutils [None req-cb805937-3829-46aa-b450-89395080ba46 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Acquired lock "refresh_cache-4ed90d16-81a6-4dbd-8936-0e137151171f" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.804773] env[62460]: DEBUG nova.network.neutron [None req-cb805937-3829-46aa-b450-89395080ba46 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 4ed90d16-81a6-4dbd-8936-0e137151171f] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 780.973826] env[62460]: DEBUG oslo_concurrency.lockutils [None req-f9eb1e2f-5ded-4741-b2a2-0483b1e8e0b6 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Lock "be50b542-4c61-4da3-af4e-8a57e05a77ee" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 114.716s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.075361] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.799s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.076082] env[62460]: ERROR nova.compute.manager [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 44c1e29f-674f-41c8-9ce8-737c88692e0b, please check neutron logs for more information. [ 781.076082] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] Traceback (most recent call last): [ 781.076082] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 781.076082] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] self.driver.spawn(context, instance, image_meta, [ 781.076082] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 781.076082] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] self._vmops.spawn(context, instance, image_meta, injected_files, [ 781.076082] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 781.076082] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] vm_ref = self.build_virtual_machine(instance, [ 781.076082] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 781.076082] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] vif_infos = vmwarevif.get_vif_info(self._session, [ 781.076082] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 781.076537] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] for vif in network_info: [ 781.076537] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 781.076537] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] return self._sync_wrapper(fn, *args, **kwargs) [ 781.076537] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 781.076537] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] self.wait() [ 781.076537] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 781.076537] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] self[:] = self._gt.wait() [ 781.076537] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 781.076537] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] return self._exit_event.wait() [ 781.076537] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 781.076537] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] current.throw(*self._exc) [ 781.076537] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 781.076537] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] result = function(*args, **kwargs) [ 781.076870] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 781.076870] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] return func(*args, **kwargs) [ 781.076870] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 781.076870] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] raise e [ 781.076870] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 781.076870] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] nwinfo = self.network_api.allocate_for_instance( [ 781.076870] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 781.076870] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] created_port_ids = self._update_ports_for_instance( [ 781.076870] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 781.076870] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] with excutils.save_and_reraise_exception(): [ 781.076870] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 781.076870] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] self.force_reraise() [ 781.076870] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 781.077322] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] raise self.value [ 781.077322] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 781.077322] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] updated_port = self._update_port( [ 781.077322] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 781.077322] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] _ensure_no_port_binding_failure(port) [ 781.077322] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 781.077322] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] raise exception.PortBindingFailed(port_id=port['id']) [ 781.077322] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] nova.exception.PortBindingFailed: Binding failed for port 44c1e29f-674f-41c8-9ce8-737c88692e0b, please check neutron logs for more information. [ 781.077322] env[62460]: ERROR nova.compute.manager [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] [ 781.077322] env[62460]: DEBUG nova.compute.utils [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] Binding failed for port 44c1e29f-674f-41c8-9ce8-737c88692e0b, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 781.078405] env[62460]: DEBUG nova.compute.manager [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] Build of instance 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe was re-scheduled: Binding failed for port 44c1e29f-674f-41c8-9ce8-737c88692e0b, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 781.078827] env[62460]: DEBUG nova.compute.manager [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 781.079063] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] Acquiring lock "refresh_cache-4f3c7bd0-314a-45d2-a75d-a22b29d67cfe" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.079216] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] Acquired lock "refresh_cache-4f3c7bd0-314a-45d2-a75d-a22b29d67cfe" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.079377] env[62460]: DEBUG nova.network.neutron [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 781.080380] env[62460]: DEBUG oslo_concurrency.lockutils [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.753s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.084302] env[62460]: INFO nova.compute.claims [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 781.321780] env[62460]: DEBUG nova.network.neutron [None req-cb805937-3829-46aa-b450-89395080ba46 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 4ed90d16-81a6-4dbd-8936-0e137151171f] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 781.381134] env[62460]: DEBUG nova.network.neutron [None req-cb805937-3829-46aa-b450-89395080ba46 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 4ed90d16-81a6-4dbd-8936-0e137151171f] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.478026] env[62460]: DEBUG nova.compute.manager [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 781.603876] env[62460]: DEBUG nova.network.neutron [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 781.698677] env[62460]: DEBUG nova.network.neutron [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.885682] env[62460]: DEBUG oslo_concurrency.lockutils [None req-cb805937-3829-46aa-b450-89395080ba46 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Releasing lock "refresh_cache-4ed90d16-81a6-4dbd-8936-0e137151171f" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.886143] env[62460]: DEBUG nova.compute.manager [None req-cb805937-3829-46aa-b450-89395080ba46 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 4ed90d16-81a6-4dbd-8936-0e137151171f] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 781.886348] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-cb805937-3829-46aa-b450-89395080ba46 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 4ed90d16-81a6-4dbd-8936-0e137151171f] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 781.887263] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7110afa-c17d-4129-91b4-3b7bfd992c20 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.894803] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb805937-3829-46aa-b450-89395080ba46 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 4ed90d16-81a6-4dbd-8936-0e137151171f] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 781.895047] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f12c6eda-950a-4487-a27f-15ee69b22c44 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.901047] env[62460]: DEBUG oslo_vmware.api [None req-cb805937-3829-46aa-b450-89395080ba46 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Waiting for the task: (returnval){ [ 781.901047] env[62460]: value = "task-1313475" [ 781.901047] env[62460]: _type = "Task" [ 781.901047] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.908322] env[62460]: DEBUG oslo_vmware.api [None req-cb805937-3829-46aa-b450-89395080ba46 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': task-1313475, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.998084] env[62460]: DEBUG oslo_concurrency.lockutils [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 782.184575] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Acquiring lock "a02b4be1-91b0-4254-8d60-654885e24f6b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 782.184868] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Lock "a02b4be1-91b0-4254-8d60-654885e24f6b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 782.204846] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] Releasing lock "refresh_cache-4f3c7bd0-314a-45d2-a75d-a22b29d67cfe" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.205091] env[62460]: DEBUG nova.compute.manager [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 782.205310] env[62460]: DEBUG nova.compute.manager [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 782.205508] env[62460]: DEBUG nova.network.neutron [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 782.220531] env[62460]: DEBUG nova.network.neutron [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 782.331471] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d59847cb-1d3d-42fd-9717-df72fa825485 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.340135] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-169b5730-ad1a-424b-90a4-3d9fe18d9c7f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.368676] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a9c0559-dedb-44bc-962c-3cff2576b2d3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.375480] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88cee2a3-b0b7-4c21-9af1-45ccc5b0f16c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.387985] env[62460]: DEBUG nova.compute.provider_tree [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 782.410605] env[62460]: DEBUG oslo_vmware.api [None req-cb805937-3829-46aa-b450-89395080ba46 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': task-1313475, 'name': PowerOffVM_Task, 'duration_secs': 0.129577} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.410862] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb805937-3829-46aa-b450-89395080ba46 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 4ed90d16-81a6-4dbd-8936-0e137151171f] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 782.411044] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-cb805937-3829-46aa-b450-89395080ba46 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 4ed90d16-81a6-4dbd-8936-0e137151171f] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 782.411282] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1d4b92a1-a36a-4c90-b97b-f098efc304e5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.437842] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-cb805937-3829-46aa-b450-89395080ba46 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 4ed90d16-81a6-4dbd-8936-0e137151171f] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 782.438080] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-cb805937-3829-46aa-b450-89395080ba46 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 4ed90d16-81a6-4dbd-8936-0e137151171f] Deleting contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 782.438277] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-cb805937-3829-46aa-b450-89395080ba46 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Deleting the datastore file [datastore1] 4ed90d16-81a6-4dbd-8936-0e137151171f {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 782.438526] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8bc1b42f-a806-4380-be7b-a5cdd0e73c7c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.445029] env[62460]: DEBUG oslo_vmware.api [None req-cb805937-3829-46aa-b450-89395080ba46 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Waiting for the task: (returnval){ [ 782.445029] env[62460]: value = "task-1313477" [ 782.445029] env[62460]: _type = "Task" [ 782.445029] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.452374] env[62460]: DEBUG oslo_vmware.api [None req-cb805937-3829-46aa-b450-89395080ba46 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': task-1313477, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.723439] env[62460]: DEBUG nova.network.neutron [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 782.894484] env[62460]: DEBUG nova.scheduler.client.report [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 782.955073] env[62460]: DEBUG oslo_vmware.api [None req-cb805937-3829-46aa-b450-89395080ba46 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Task: {'id': task-1313477, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.089736} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.955342] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-cb805937-3829-46aa-b450-89395080ba46 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 782.955617] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-cb805937-3829-46aa-b450-89395080ba46 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 4ed90d16-81a6-4dbd-8936-0e137151171f] Deleted contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 782.955700] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-cb805937-3829-46aa-b450-89395080ba46 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 4ed90d16-81a6-4dbd-8936-0e137151171f] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 782.955955] env[62460]: INFO nova.compute.manager [None req-cb805937-3829-46aa-b450-89395080ba46 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] [instance: 4ed90d16-81a6-4dbd-8936-0e137151171f] Took 1.07 seconds to destroy the instance on the hypervisor. [ 782.956126] env[62460]: DEBUG oslo.service.loopingcall [None req-cb805937-3829-46aa-b450-89395080ba46 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 782.956327] env[62460]: DEBUG nova.compute.manager [-] [instance: 4ed90d16-81a6-4dbd-8936-0e137151171f] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 782.956486] env[62460]: DEBUG nova.network.neutron [-] [instance: 4ed90d16-81a6-4dbd-8936-0e137151171f] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 782.971485] env[62460]: DEBUG nova.network.neutron [-] [instance: 4ed90d16-81a6-4dbd-8936-0e137151171f] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 783.226618] env[62460]: INFO nova.compute.manager [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] [instance: 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe] Took 1.02 seconds to deallocate network for instance. [ 783.400220] env[62460]: DEBUG oslo_concurrency.lockutils [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.320s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.400878] env[62460]: DEBUG nova.compute.manager [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 783.403704] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.185s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 783.405023] env[62460]: INFO nova.compute.claims [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 783.474629] env[62460]: DEBUG nova.network.neutron [-] [instance: 4ed90d16-81a6-4dbd-8936-0e137151171f] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.909886] env[62460]: DEBUG nova.compute.utils [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 783.914064] env[62460]: DEBUG nova.compute.manager [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 783.914334] env[62460]: DEBUG nova.network.neutron [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 783.962737] env[62460]: DEBUG nova.policy [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd7ac705ce4dc4f42b89cf70ac3316f71', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '335cfe76a8c44686891e17906c96d158', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 783.977164] env[62460]: INFO nova.compute.manager [-] [instance: 4ed90d16-81a6-4dbd-8936-0e137151171f] Took 1.02 seconds to deallocate network for instance. [ 784.258050] env[62460]: DEBUG nova.network.neutron [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] Successfully created port: d8f1c890-d4c1-42ce-9093-2d21538c4bec {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 784.261222] env[62460]: INFO nova.scheduler.client.report [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] Deleted allocations for instance 4f3c7bd0-314a-45d2-a75d-a22b29d67cfe [ 784.417182] env[62460]: DEBUG nova.compute.manager [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 784.482923] env[62460]: DEBUG oslo_concurrency.lockutils [None req-cb805937-3829-46aa-b450-89395080ba46 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.655584] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dcbee43-d090-4218-9d28-4880ded90785 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.663773] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7047df1-9beb-45a9-b0de-23ed753efb56 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.701072] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9cdc13f-0b79-41c6-8e80-311653bc3a14 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.708851] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81017989-ff46-40ac-87ea-d24ff4e20fe2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.722718] env[62460]: DEBUG nova.compute.provider_tree [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 784.770612] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5959e05a-6e84-4f59-a451-9a5f9efcff21 tempest-AttachInterfacesV270Test-321697551 tempest-AttachInterfacesV270Test-321697551-project-member] Lock "4f3c7bd0-314a-45d2-a75d-a22b29d67cfe" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 110.936s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.029549] env[62460]: DEBUG nova.compute.manager [req-ba94fa62-7f16-4936-b9bc-722d92a17670 req-48121e28-9378-48f1-823f-6a9d52066acf service nova] [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] Received event network-changed-d8f1c890-d4c1-42ce-9093-2d21538c4bec {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 785.029801] env[62460]: DEBUG nova.compute.manager [req-ba94fa62-7f16-4936-b9bc-722d92a17670 req-48121e28-9378-48f1-823f-6a9d52066acf service nova] [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] Refreshing instance network info cache due to event network-changed-d8f1c890-d4c1-42ce-9093-2d21538c4bec. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 785.030035] env[62460]: DEBUG oslo_concurrency.lockutils [req-ba94fa62-7f16-4936-b9bc-722d92a17670 req-48121e28-9378-48f1-823f-6a9d52066acf service nova] Acquiring lock "refresh_cache-c7d51b39-b449-4af3-a4d0-c746983ded3e" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 785.030406] env[62460]: DEBUG oslo_concurrency.lockutils [req-ba94fa62-7f16-4936-b9bc-722d92a17670 req-48121e28-9378-48f1-823f-6a9d52066acf service nova] Acquired lock "refresh_cache-c7d51b39-b449-4af3-a4d0-c746983ded3e" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 785.030588] env[62460]: DEBUG nova.network.neutron [req-ba94fa62-7f16-4936-b9bc-722d92a17670 req-48121e28-9378-48f1-823f-6a9d52066acf service nova] [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] Refreshing network info cache for port d8f1c890-d4c1-42ce-9093-2d21538c4bec {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 785.226276] env[62460]: DEBUG nova.scheduler.client.report [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 785.274407] env[62460]: DEBUG nova.compute.manager [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 785.291090] env[62460]: ERROR nova.compute.manager [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d8f1c890-d4c1-42ce-9093-2d21538c4bec, please check neutron logs for more information. [ 785.291090] env[62460]: ERROR nova.compute.manager Traceback (most recent call last): [ 785.291090] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 785.291090] env[62460]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 785.291090] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 785.291090] env[62460]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 785.291090] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 785.291090] env[62460]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 785.291090] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 785.291090] env[62460]: ERROR nova.compute.manager self.force_reraise() [ 785.291090] env[62460]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 785.291090] env[62460]: ERROR nova.compute.manager raise self.value [ 785.291090] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 785.291090] env[62460]: ERROR nova.compute.manager updated_port = self._update_port( [ 785.291090] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 785.291090] env[62460]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 785.291554] env[62460]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 785.291554] env[62460]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 785.291554] env[62460]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d8f1c890-d4c1-42ce-9093-2d21538c4bec, please check neutron logs for more information. [ 785.291554] env[62460]: ERROR nova.compute.manager [ 785.291554] env[62460]: Traceback (most recent call last): [ 785.291554] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 785.291554] env[62460]: listener.cb(fileno) [ 785.291554] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 785.291554] env[62460]: result = function(*args, **kwargs) [ 785.291554] env[62460]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 785.291554] env[62460]: return func(*args, **kwargs) [ 785.291554] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 785.291554] env[62460]: raise e [ 785.291554] env[62460]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 785.291554] env[62460]: nwinfo = self.network_api.allocate_for_instance( [ 785.291554] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 785.291554] env[62460]: created_port_ids = self._update_ports_for_instance( [ 785.291554] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 785.291554] env[62460]: with excutils.save_and_reraise_exception(): [ 785.291554] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 785.291554] env[62460]: self.force_reraise() [ 785.291554] env[62460]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 785.291554] env[62460]: raise self.value [ 785.291554] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 785.291554] env[62460]: updated_port = self._update_port( [ 785.291554] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 785.291554] env[62460]: _ensure_no_port_binding_failure(port) [ 785.291554] env[62460]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 785.291554] env[62460]: raise exception.PortBindingFailed(port_id=port['id']) [ 785.292362] env[62460]: nova.exception.PortBindingFailed: Binding failed for port d8f1c890-d4c1-42ce-9093-2d21538c4bec, please check neutron logs for more information. [ 785.292362] env[62460]: Removing descriptor: 18 [ 785.427070] env[62460]: DEBUG nova.compute.manager [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 785.451988] env[62460]: DEBUG nova.virt.hardware [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 785.452247] env[62460]: DEBUG nova.virt.hardware [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 785.452405] env[62460]: DEBUG nova.virt.hardware [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 785.452585] env[62460]: DEBUG nova.virt.hardware [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 785.452731] env[62460]: DEBUG nova.virt.hardware [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 785.452889] env[62460]: DEBUG nova.virt.hardware [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 785.453108] env[62460]: DEBUG nova.virt.hardware [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 785.453270] env[62460]: DEBUG nova.virt.hardware [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 785.453433] env[62460]: DEBUG nova.virt.hardware [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 785.453594] env[62460]: DEBUG nova.virt.hardware [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 785.453763] env[62460]: DEBUG nova.virt.hardware [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 785.454658] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d93377f6-7bf5-4ad2-a739-2db4bbc8143a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.463350] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8404b06b-b93e-47ef-8250-2c7ba8b06e74 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.477055] env[62460]: ERROR nova.compute.manager [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d8f1c890-d4c1-42ce-9093-2d21538c4bec, please check neutron logs for more information. [ 785.477055] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] Traceback (most recent call last): [ 785.477055] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 785.477055] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] yield resources [ 785.477055] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 785.477055] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] self.driver.spawn(context, instance, image_meta, [ 785.477055] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 785.477055] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 785.477055] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 785.477055] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] vm_ref = self.build_virtual_machine(instance, [ 785.477055] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 785.477402] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] vif_infos = vmwarevif.get_vif_info(self._session, [ 785.477402] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 785.477402] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] for vif in network_info: [ 785.477402] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 785.477402] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] return self._sync_wrapper(fn, *args, **kwargs) [ 785.477402] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 785.477402] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] self.wait() [ 785.477402] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 785.477402] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] self[:] = self._gt.wait() [ 785.477402] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 785.477402] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] return self._exit_event.wait() [ 785.477402] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 785.477402] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] current.throw(*self._exc) [ 785.477677] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 785.477677] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] result = function(*args, **kwargs) [ 785.477677] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 785.477677] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] return func(*args, **kwargs) [ 785.477677] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 785.477677] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] raise e [ 785.477677] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 785.477677] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] nwinfo = self.network_api.allocate_for_instance( [ 785.477677] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 785.477677] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] created_port_ids = self._update_ports_for_instance( [ 785.477677] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 785.477677] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] with excutils.save_and_reraise_exception(): [ 785.477677] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 785.477960] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] self.force_reraise() [ 785.477960] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 785.477960] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] raise self.value [ 785.477960] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 785.477960] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] updated_port = self._update_port( [ 785.477960] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 785.477960] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] _ensure_no_port_binding_failure(port) [ 785.477960] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 785.477960] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] raise exception.PortBindingFailed(port_id=port['id']) [ 785.477960] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] nova.exception.PortBindingFailed: Binding failed for port d8f1c890-d4c1-42ce-9093-2d21538c4bec, please check neutron logs for more information. [ 785.477960] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] [ 785.477960] env[62460]: INFO nova.compute.manager [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] Terminating instance [ 785.483800] env[62460]: DEBUG oslo_concurrency.lockutils [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Acquiring lock "refresh_cache-c7d51b39-b449-4af3-a4d0-c746983ded3e" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 785.549054] env[62460]: DEBUG nova.network.neutron [req-ba94fa62-7f16-4936-b9bc-722d92a17670 req-48121e28-9378-48f1-823f-6a9d52066acf service nova] [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 785.629284] env[62460]: DEBUG nova.network.neutron [req-ba94fa62-7f16-4936-b9bc-722d92a17670 req-48121e28-9378-48f1-823f-6a9d52066acf service nova] [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.734540] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.331s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.735230] env[62460]: DEBUG nova.compute.manager [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 785.739713] env[62460]: DEBUG oslo_concurrency.lockutils [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.820s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.803974] env[62460]: DEBUG oslo_concurrency.lockutils [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.131760] env[62460]: DEBUG oslo_concurrency.lockutils [req-ba94fa62-7f16-4936-b9bc-722d92a17670 req-48121e28-9378-48f1-823f-6a9d52066acf service nova] Releasing lock "refresh_cache-c7d51b39-b449-4af3-a4d0-c746983ded3e" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 786.132146] env[62460]: DEBUG oslo_concurrency.lockutils [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Acquired lock "refresh_cache-c7d51b39-b449-4af3-a4d0-c746983ded3e" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.132343] env[62460]: DEBUG nova.network.neutron [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 786.247695] env[62460]: DEBUG nova.compute.utils [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 786.249690] env[62460]: DEBUG nova.compute.manager [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Not allocating networking since 'none' was specified. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 786.487746] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68ce1917-c556-4f48-895c-e2861019a69a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.496017] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61d72b9a-fa2f-4b6a-b6b1-4e100047fa9c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.525425] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0af3ad1a-948e-49e9-917e-8bd1192a4dcb {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.532703] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8339678b-77d0-4105-af92-1865afa2b0ed {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.545615] env[62460]: DEBUG nova.compute.provider_tree [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 786.653600] env[62460]: DEBUG nova.network.neutron [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 786.740577] env[62460]: DEBUG nova.network.neutron [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 786.751030] env[62460]: DEBUG nova.compute.manager [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 787.051230] env[62460]: DEBUG nova.scheduler.client.report [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 787.087692] env[62460]: DEBUG nova.compute.manager [req-918e68ad-ef40-4838-b71c-432df3ef2279 req-ff7d2858-2c65-4f72-8849-a1e7f9ee0e0f service nova] [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] Received event network-vif-deleted-d8f1c890-d4c1-42ce-9093-2d21538c4bec {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 787.243172] env[62460]: DEBUG oslo_concurrency.lockutils [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Releasing lock "refresh_cache-c7d51b39-b449-4af3-a4d0-c746983ded3e" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 787.243587] env[62460]: DEBUG nova.compute.manager [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 787.243783] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 787.244122] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2c7f2ac8-9d9e-476e-8b8a-6a28ec4e4511 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.256868] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63c5900e-3b57-4a92-a90f-368509faf3fd {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.286984] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c7d51b39-b449-4af3-a4d0-c746983ded3e could not be found. [ 787.288140] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 787.288140] env[62460]: INFO nova.compute.manager [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] Took 0.04 seconds to destroy the instance on the hypervisor. [ 787.288140] env[62460]: DEBUG oslo.service.loopingcall [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 787.289095] env[62460]: DEBUG nova.compute.manager [-] [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 787.289460] env[62460]: DEBUG nova.network.neutron [-] [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 787.312536] env[62460]: DEBUG nova.network.neutron [-] [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 787.558017] env[62460]: DEBUG oslo_concurrency.lockutils [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.816s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.558017] env[62460]: ERROR nova.compute.manager [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0fcaca39-5cf6-4276-9963-5a22a97e131f, please check neutron logs for more information. [ 787.558017] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] Traceback (most recent call last): [ 787.558017] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 787.558017] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] self.driver.spawn(context, instance, image_meta, [ 787.558017] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 787.558017] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 787.558017] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 787.558017] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] vm_ref = self.build_virtual_machine(instance, [ 787.558314] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 787.558314] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] vif_infos = vmwarevif.get_vif_info(self._session, [ 787.558314] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 787.558314] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] for vif in network_info: [ 787.558314] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 787.558314] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] return self._sync_wrapper(fn, *args, **kwargs) [ 787.558314] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 787.558314] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] self.wait() [ 787.558314] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 787.558314] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] self[:] = self._gt.wait() [ 787.558314] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 787.558314] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] return self._exit_event.wait() [ 787.558314] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 787.558655] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] current.throw(*self._exc) [ 787.558655] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 787.558655] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] result = function(*args, **kwargs) [ 787.558655] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 787.558655] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] return func(*args, **kwargs) [ 787.558655] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 787.558655] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] raise e [ 787.558655] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 787.558655] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] nwinfo = self.network_api.allocate_for_instance( [ 787.558655] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 787.558655] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] created_port_ids = self._update_ports_for_instance( [ 787.558655] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 787.558655] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] with excutils.save_and_reraise_exception(): [ 787.559011] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 787.559011] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] self.force_reraise() [ 787.559011] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 787.559011] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] raise self.value [ 787.559011] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 787.559011] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] updated_port = self._update_port( [ 787.559011] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 787.559011] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] _ensure_no_port_binding_failure(port) [ 787.559011] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 787.559011] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] raise exception.PortBindingFailed(port_id=port['id']) [ 787.559011] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] nova.exception.PortBindingFailed: Binding failed for port 0fcaca39-5cf6-4276-9963-5a22a97e131f, please check neutron logs for more information. [ 787.559011] env[62460]: ERROR nova.compute.manager [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] [ 787.559327] env[62460]: DEBUG nova.compute.utils [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] Binding failed for port 0fcaca39-5cf6-4276-9963-5a22a97e131f, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 787.559592] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.946s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.561259] env[62460]: INFO nova.compute.claims [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 787.566016] env[62460]: DEBUG nova.compute.manager [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] Build of instance 559757ef-ee92-4b88-8631-6f743fb88bc6 was re-scheduled: Binding failed for port 0fcaca39-5cf6-4276-9963-5a22a97e131f, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 787.566016] env[62460]: DEBUG nova.compute.manager [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 787.566016] env[62460]: DEBUG oslo_concurrency.lockutils [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] Acquiring lock "refresh_cache-559757ef-ee92-4b88-8631-6f743fb88bc6" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.566016] env[62460]: DEBUG oslo_concurrency.lockutils [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] Acquired lock "refresh_cache-559757ef-ee92-4b88-8631-6f743fb88bc6" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.566198] env[62460]: DEBUG nova.network.neutron [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 787.774946] env[62460]: DEBUG nova.compute.manager [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 787.802983] env[62460]: DEBUG nova.virt.hardware [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 787.807477] env[62460]: DEBUG nova.virt.hardware [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 787.807669] env[62460]: DEBUG nova.virt.hardware [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 787.809110] env[62460]: DEBUG nova.virt.hardware [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 787.809293] env[62460]: DEBUG nova.virt.hardware [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 787.809517] env[62460]: DEBUG nova.virt.hardware [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 787.809970] env[62460]: DEBUG nova.virt.hardware [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 787.810167] env[62460]: DEBUG nova.virt.hardware [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 787.810348] env[62460]: DEBUG nova.virt.hardware [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 787.811719] env[62460]: DEBUG nova.virt.hardware [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 787.811955] env[62460]: DEBUG nova.virt.hardware [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 787.815113] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cea52aea-2838-417e-881f-d74ddf200c48 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.820141] env[62460]: DEBUG nova.network.neutron [-] [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 787.827341] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5450c038-6752-4107-9939-4a5a4a5f1059 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.841815] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Instance VIF info [] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 787.847436] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Creating folder: Project (d18eb4c53ed740bd904403498980bc0a). Parent ref: group-v281134. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 787.847997] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-21e9901e-ee5a-4373-bb7d-2f592b2a56e8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.858777] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Created folder: Project (d18eb4c53ed740bd904403498980bc0a) in parent group-v281134. [ 787.858976] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Creating folder: Instances. Parent ref: group-v281155. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 787.862866] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-39dbfe3a-4aa5-4439-8db1-0eb48474c9d2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.868467] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Created folder: Instances in parent group-v281155. [ 787.868750] env[62460]: DEBUG oslo.service.loopingcall [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 787.868947] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 787.869163] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-32ed10d4-8605-495d-aef8-47e8cd2d8aa0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.886690] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 787.886690] env[62460]: value = "task-1313480" [ 787.886690] env[62460]: _type = "Task" [ 787.886690] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.894474] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313480, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.092762] env[62460]: DEBUG nova.network.neutron [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 788.206144] env[62460]: DEBUG nova.network.neutron [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.322628] env[62460]: INFO nova.compute.manager [-] [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] Took 1.03 seconds to deallocate network for instance. [ 788.324981] env[62460]: DEBUG nova.compute.claims [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] Aborting claim: {{(pid=62460) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 788.325245] env[62460]: DEBUG oslo_concurrency.lockutils [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.397096] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313480, 'name': CreateVM_Task, 'duration_secs': 0.257048} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.397096] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 788.397502] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.397719] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.398089] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 788.398380] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d880e997-eea3-436c-ac84-6c5dd98843ee {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.403085] env[62460]: DEBUG oslo_vmware.api [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Waiting for the task: (returnval){ [ 788.403085] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52851524-4bfc-0e5e-e621-2ff912f81b52" [ 788.403085] env[62460]: _type = "Task" [ 788.403085] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.412567] env[62460]: DEBUG oslo_vmware.api [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52851524-4bfc-0e5e-e621-2ff912f81b52, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.710534] env[62460]: DEBUG oslo_concurrency.lockutils [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] Releasing lock "refresh_cache-559757ef-ee92-4b88-8631-6f743fb88bc6" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.710534] env[62460]: DEBUG nova.compute.manager [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 788.710534] env[62460]: DEBUG nova.compute.manager [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 788.710534] env[62460]: DEBUG nova.network.neutron [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 788.726596] env[62460]: DEBUG nova.network.neutron [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 788.871595] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9805814b-e786-46fe-ab71-acb000cec604 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.879763] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d147fe14-f8ed-455c-95a7-8be45bae29f3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.913561] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc2aecab-8aa7-422d-94ff-e8ecdf673871 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.922162] env[62460]: DEBUG oslo_vmware.api [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52851524-4bfc-0e5e-e621-2ff912f81b52, 'name': SearchDatastore_Task, 'duration_secs': 0.010109} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.924792] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.925051] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 788.925333] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.925489] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.925672] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 788.925969] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-acb84952-4dfd-4eb6-af1f-9c3769b8409f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.929753] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36c50363-8532-495b-abca-6601a747127d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.943984] env[62460]: DEBUG nova.compute.provider_tree [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 788.946918] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 788.947110] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 788.948034] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3fa26181-8f6d-449b-9e5b-e1705d6c49d1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.954028] env[62460]: DEBUG oslo_vmware.api [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Waiting for the task: (returnval){ [ 788.954028] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52a705f4-bb28-239c-b7f2-3a583a83ff5f" [ 788.954028] env[62460]: _type = "Task" [ 788.954028] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.965713] env[62460]: DEBUG oslo_vmware.api [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52a705f4-bb28-239c-b7f2-3a583a83ff5f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.231727] env[62460]: DEBUG nova.network.neutron [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.449352] env[62460]: DEBUG nova.scheduler.client.report [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 789.465381] env[62460]: DEBUG oslo_vmware.api [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52a705f4-bb28-239c-b7f2-3a583a83ff5f, 'name': SearchDatastore_Task, 'duration_secs': 0.008899} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.466254] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-378aa5ea-f30b-4f19-8688-8168f9b379f5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.471693] env[62460]: DEBUG oslo_vmware.api [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Waiting for the task: (returnval){ [ 789.471693] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52c85fa1-f75c-eb6e-9b77-6444499b806a" [ 789.471693] env[62460]: _type = "Task" [ 789.471693] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.481049] env[62460]: DEBUG oslo_vmware.api [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52c85fa1-f75c-eb6e-9b77-6444499b806a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.733754] env[62460]: INFO nova.compute.manager [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] [instance: 559757ef-ee92-4b88-8631-6f743fb88bc6] Took 1.02 seconds to deallocate network for instance. [ 789.960039] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.402s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.961687] env[62460]: DEBUG nova.compute.manager [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 789.972677] env[62460]: DEBUG oslo_concurrency.lockutils [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.206s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 789.972677] env[62460]: INFO nova.compute.claims [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 789.984758] env[62460]: DEBUG oslo_vmware.api [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52c85fa1-f75c-eb6e-9b77-6444499b806a, 'name': SearchDatastore_Task, 'duration_secs': 0.009368} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.985113] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.985565] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 51718896-f5bf-43a9-9396-1ac768737ba2/51718896-f5bf-43a9-9396-1ac768737ba2.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 789.985872] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ec778941-6ff2-4afe-bb14-acac6843c4bd {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.993095] env[62460]: DEBUG oslo_vmware.api [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Waiting for the task: (returnval){ [ 789.993095] env[62460]: value = "task-1313481" [ 789.993095] env[62460]: _type = "Task" [ 789.993095] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.004016] env[62460]: DEBUG oslo_vmware.api [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Task: {'id': task-1313481, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.472250] env[62460]: DEBUG nova.compute.utils [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 790.478076] env[62460]: DEBUG nova.compute.manager [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 790.478316] env[62460]: DEBUG nova.network.neutron [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 790.506144] env[62460]: DEBUG oslo_vmware.api [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Task: {'id': task-1313481, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.4658} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.506570] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 51718896-f5bf-43a9-9396-1ac768737ba2/51718896-f5bf-43a9-9396-1ac768737ba2.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 790.507729] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 790.507729] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-477ead7b-9105-4423-ad93-591a0a78e510 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.516117] env[62460]: DEBUG oslo_vmware.api [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Waiting for the task: (returnval){ [ 790.516117] env[62460]: value = "task-1313485" [ 790.516117] env[62460]: _type = "Task" [ 790.516117] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.525297] env[62460]: DEBUG oslo_vmware.api [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Task: {'id': task-1313485, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.542794] env[62460]: DEBUG nova.policy [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c1f21a1b9cd74068b066fbbe631d0c3f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f0a7a688391c4fbe9e1088c5047f00d6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 790.778231] env[62460]: INFO nova.scheduler.client.report [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] Deleted allocations for instance 559757ef-ee92-4b88-8631-6f743fb88bc6 [ 790.877799] env[62460]: DEBUG nova.network.neutron [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Successfully created port: 51efbac7-c441-4555-9a3c-d42167d75fcc {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 790.982125] env[62460]: DEBUG nova.compute.manager [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 791.025905] env[62460]: DEBUG oslo_vmware.api [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Task: {'id': task-1313485, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068831} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.029685] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 791.031319] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a459c6c-08c5-460d-a63d-3989ada376c0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.052820] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Reconfiguring VM instance instance-00000039 to attach disk [datastore1] 51718896-f5bf-43a9-9396-1ac768737ba2/51718896-f5bf-43a9-9396-1ac768737ba2.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 791.055619] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-83d822ee-f64b-4973-a09c-b4736ecdb0bb {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.075273] env[62460]: DEBUG oslo_vmware.api [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Waiting for the task: (returnval){ [ 791.075273] env[62460]: value = "task-1313486" [ 791.075273] env[62460]: _type = "Task" [ 791.075273] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.084016] env[62460]: DEBUG oslo_vmware.api [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Task: {'id': task-1313486, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.292442] env[62460]: DEBUG oslo_concurrency.lockutils [None req-038a4c0b-3831-4118-969b-8d05e0238abb tempest-ServerExternalEventsTest-1993612175 tempest-ServerExternalEventsTest-1993612175-project-member] Lock "559757ef-ee92-4b88-8631-6f743fb88bc6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 114.454s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 791.307593] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fabcf047-f63d-4182-8e4b-e1aaa9c5b42c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.316431] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ee6688a-f454-48bc-9942-5bf458b4aa4f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.345779] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ca61b94-3c42-445f-8888-628e147e305f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.353021] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f078190f-d3e0-4970-8a18-64bbd2fa4363 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.366378] env[62460]: DEBUG nova.compute.provider_tree [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 791.587201] env[62460]: DEBUG oslo_vmware.api [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Task: {'id': task-1313486, 'name': ReconfigVM_Task, 'duration_secs': 0.278646} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.587544] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Reconfigured VM instance instance-00000039 to attach disk [datastore1] 51718896-f5bf-43a9-9396-1ac768737ba2/51718896-f5bf-43a9-9396-1ac768737ba2.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 791.588187] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1fe220b7-20bb-4615-9a37-30f3b9a726e2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.595386] env[62460]: DEBUG oslo_vmware.api [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Waiting for the task: (returnval){ [ 791.595386] env[62460]: value = "task-1313487" [ 791.595386] env[62460]: _type = "Task" [ 791.595386] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.606667] env[62460]: DEBUG oslo_vmware.api [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Task: {'id': task-1313487, 'name': Rename_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.796547] env[62460]: DEBUG nova.compute.manager [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 791.871722] env[62460]: DEBUG nova.scheduler.client.report [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 791.993058] env[62460]: DEBUG nova.compute.manager [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 792.017026] env[62460]: DEBUG nova.virt.hardware [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 792.017286] env[62460]: DEBUG nova.virt.hardware [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 792.017453] env[62460]: DEBUG nova.virt.hardware [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 792.017634] env[62460]: DEBUG nova.virt.hardware [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 792.017778] env[62460]: DEBUG nova.virt.hardware [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 792.017936] env[62460]: DEBUG nova.virt.hardware [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 792.018145] env[62460]: DEBUG nova.virt.hardware [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 792.018566] env[62460]: DEBUG nova.virt.hardware [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 792.018566] env[62460]: DEBUG nova.virt.hardware [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 792.018660] env[62460]: DEBUG nova.virt.hardware [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 792.018821] env[62460]: DEBUG nova.virt.hardware [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 792.019955] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78895e6c-17bb-4e0c-b122-2635ca556d0a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.028535] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33dc4d17-ca95-4a6b-b82e-161396ec5401 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.104118] env[62460]: DEBUG oslo_vmware.api [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Task: {'id': task-1313487, 'name': Rename_Task, 'duration_secs': 0.135991} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.104460] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 792.104717] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bf1b876f-ef89-4392-b911-7a48037a447d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.110833] env[62460]: DEBUG oslo_vmware.api [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Waiting for the task: (returnval){ [ 792.110833] env[62460]: value = "task-1313488" [ 792.110833] env[62460]: _type = "Task" [ 792.110833] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.118610] env[62460]: DEBUG oslo_vmware.api [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Task: {'id': task-1313488, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.321932] env[62460]: DEBUG oslo_concurrency.lockutils [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.377172] env[62460]: DEBUG oslo_concurrency.lockutils [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.408s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.377696] env[62460]: DEBUG nova.compute.manager [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 792.380576] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.117s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.538541] env[62460]: DEBUG nova.compute.manager [req-95db536d-6641-4aa9-aeb3-9071c19a1d4e req-5b901d55-514c-4875-962a-762ae23eb4ad service nova] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Received event network-vif-plugged-51efbac7-c441-4555-9a3c-d42167d75fcc {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 792.538775] env[62460]: DEBUG oslo_concurrency.lockutils [req-95db536d-6641-4aa9-aeb3-9071c19a1d4e req-5b901d55-514c-4875-962a-762ae23eb4ad service nova] Acquiring lock "1f318a64-2c38-470b-8fae-4ba4543a5681-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.538991] env[62460]: DEBUG oslo_concurrency.lockutils [req-95db536d-6641-4aa9-aeb3-9071c19a1d4e req-5b901d55-514c-4875-962a-762ae23eb4ad service nova] Lock "1f318a64-2c38-470b-8fae-4ba4543a5681-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.539664] env[62460]: DEBUG oslo_concurrency.lockutils [req-95db536d-6641-4aa9-aeb3-9071c19a1d4e req-5b901d55-514c-4875-962a-762ae23eb4ad service nova] Lock "1f318a64-2c38-470b-8fae-4ba4543a5681-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.539860] env[62460]: DEBUG nova.compute.manager [req-95db536d-6641-4aa9-aeb3-9071c19a1d4e req-5b901d55-514c-4875-962a-762ae23eb4ad service nova] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] No waiting events found dispatching network-vif-plugged-51efbac7-c441-4555-9a3c-d42167d75fcc {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 792.540048] env[62460]: WARNING nova.compute.manager [req-95db536d-6641-4aa9-aeb3-9071c19a1d4e req-5b901d55-514c-4875-962a-762ae23eb4ad service nova] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Received unexpected event network-vif-plugged-51efbac7-c441-4555-9a3c-d42167d75fcc for instance with vm_state building and task_state spawning. [ 792.625022] env[62460]: DEBUG oslo_vmware.api [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Task: {'id': task-1313488, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.713151] env[62460]: DEBUG nova.network.neutron [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Successfully updated port: 51efbac7-c441-4555-9a3c-d42167d75fcc {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 792.889957] env[62460]: DEBUG nova.compute.utils [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 792.892064] env[62460]: DEBUG nova.compute.manager [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 792.892064] env[62460]: DEBUG nova.network.neutron [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 792.948996] env[62460]: DEBUG nova.policy [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd7b38fb0a53f4d078ae86d68e3eb6268', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '53fa06008e9f43488362895e7a143700', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 793.130856] env[62460]: DEBUG oslo_vmware.api [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Task: {'id': task-1313488, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.220403] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Acquiring lock "refresh_cache-1f318a64-2c38-470b-8fae-4ba4543a5681" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.220528] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Acquired lock "refresh_cache-1f318a64-2c38-470b-8fae-4ba4543a5681" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.220893] env[62460]: DEBUG nova.network.neutron [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 793.224392] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1939a6f0-b465-4df2-a5d3-139b1754294d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.235581] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab2b5fa3-3f2e-4634-a5c9-364d8938c4a2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.277634] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0524d5df-db9b-47df-ac37-456cf1f0baa2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.285831] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48f6cdeb-98ff-4816-8ecc-79a37a2bd42c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.302799] env[62460]: DEBUG nova.compute.provider_tree [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 793.308582] env[62460]: DEBUG nova.network.neutron [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Successfully created port: ba38cbb0-62cf-47cb-a238-87ebc03de250 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 793.396594] env[62460]: DEBUG nova.compute.manager [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 793.624562] env[62460]: DEBUG oslo_vmware.api [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Task: {'id': task-1313488, 'name': PowerOnVM_Task, 'duration_secs': 1.049257} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.624930] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 793.625172] env[62460]: INFO nova.compute.manager [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Took 5.85 seconds to spawn the instance on the hypervisor. [ 793.625430] env[62460]: DEBUG nova.compute.manager [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 793.626246] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dd9e092-3500-4dac-8152-73d0db6a1f4c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.760323] env[62460]: DEBUG nova.network.neutron [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 793.810384] env[62460]: DEBUG nova.scheduler.client.report [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 793.976837] env[62460]: DEBUG nova.network.neutron [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Updating instance_info_cache with network_info: [{"id": "51efbac7-c441-4555-9a3c-d42167d75fcc", "address": "fa:16:3e:a7:a6:af", "network": {"id": "6d99b7f4-65ae-4a8f-aa6e-bce5b845dbe1", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1836645522-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f0a7a688391c4fbe9e1088c5047f00d6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ab2e9f5-54fd-4cab-9405-ed65e2aaba64", "external-id": "nsx-vlan-transportzone-222", "segmentation_id": 222, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51efbac7-c4", "ovs_interfaceid": "51efbac7-c441-4555-9a3c-d42167d75fcc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 794.144368] env[62460]: INFO nova.compute.manager [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Took 28.95 seconds to build instance. [ 794.317940] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.937s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.318627] env[62460]: ERROR nova.compute.manager [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b4e718dc-828e-49ea-8bd6-83b0a34f6d57, please check neutron logs for more information. [ 794.318627] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] Traceback (most recent call last): [ 794.318627] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 794.318627] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] self.driver.spawn(context, instance, image_meta, [ 794.318627] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 794.318627] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 794.318627] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 794.318627] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] vm_ref = self.build_virtual_machine(instance, [ 794.318627] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 794.318627] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] vif_infos = vmwarevif.get_vif_info(self._session, [ 794.318627] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 794.318940] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] for vif in network_info: [ 794.318940] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 794.318940] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] return self._sync_wrapper(fn, *args, **kwargs) [ 794.318940] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 794.318940] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] self.wait() [ 794.318940] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 794.318940] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] self[:] = self._gt.wait() [ 794.318940] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 794.318940] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] return self._exit_event.wait() [ 794.318940] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 794.318940] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] result = hub.switch() [ 794.318940] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 794.318940] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] return self.greenlet.switch() [ 794.319273] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 794.319273] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] result = function(*args, **kwargs) [ 794.319273] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 794.319273] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] return func(*args, **kwargs) [ 794.319273] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 794.319273] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] raise e [ 794.319273] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 794.319273] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] nwinfo = self.network_api.allocate_for_instance( [ 794.319273] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 794.319273] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] created_port_ids = self._update_ports_for_instance( [ 794.319273] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 794.319273] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] with excutils.save_and_reraise_exception(): [ 794.319273] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 794.319581] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] self.force_reraise() [ 794.319581] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 794.319581] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] raise self.value [ 794.319581] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 794.319581] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] updated_port = self._update_port( [ 794.319581] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 794.319581] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] _ensure_no_port_binding_failure(port) [ 794.319581] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 794.319581] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] raise exception.PortBindingFailed(port_id=port['id']) [ 794.319581] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] nova.exception.PortBindingFailed: Binding failed for port b4e718dc-828e-49ea-8bd6-83b0a34f6d57, please check neutron logs for more information. [ 794.319581] env[62460]: ERROR nova.compute.manager [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] [ 794.319848] env[62460]: DEBUG nova.compute.utils [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] Binding failed for port b4e718dc-828e-49ea-8bd6-83b0a34f6d57, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 794.320715] env[62460]: DEBUG oslo_concurrency.lockutils [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.743s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.322211] env[62460]: INFO nova.compute.claims [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 794.324895] env[62460]: DEBUG nova.compute.manager [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] Build of instance 588e5489-2f61-473a-90e9-883eb8b163c8 was re-scheduled: Binding failed for port b4e718dc-828e-49ea-8bd6-83b0a34f6d57, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 794.325420] env[62460]: DEBUG nova.compute.manager [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 794.325694] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Acquiring lock "refresh_cache-588e5489-2f61-473a-90e9-883eb8b163c8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.325850] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Acquired lock "refresh_cache-588e5489-2f61-473a-90e9-883eb8b163c8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.326023] env[62460]: DEBUG nova.network.neutron [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 794.410092] env[62460]: DEBUG nova.compute.manager [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 794.444182] env[62460]: DEBUG nova.virt.hardware [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 794.444556] env[62460]: DEBUG nova.virt.hardware [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 794.444763] env[62460]: DEBUG nova.virt.hardware [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 794.444973] env[62460]: DEBUG nova.virt.hardware [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 794.445208] env[62460]: DEBUG nova.virt.hardware [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 794.445307] env[62460]: DEBUG nova.virt.hardware [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 794.445515] env[62460]: DEBUG nova.virt.hardware [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 794.445678] env[62460]: DEBUG nova.virt.hardware [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 794.445847] env[62460]: DEBUG nova.virt.hardware [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 794.446038] env[62460]: DEBUG nova.virt.hardware [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 794.446228] env[62460]: DEBUG nova.virt.hardware [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 794.447130] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0dbda29-68a4-459f-8d0b-88ef5caa888b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.456277] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee379326-1a0e-430f-ac0d-def225642c23 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.479360] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Releasing lock "refresh_cache-1f318a64-2c38-470b-8fae-4ba4543a5681" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.479690] env[62460]: DEBUG nova.compute.manager [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Instance network_info: |[{"id": "51efbac7-c441-4555-9a3c-d42167d75fcc", "address": "fa:16:3e:a7:a6:af", "network": {"id": "6d99b7f4-65ae-4a8f-aa6e-bce5b845dbe1", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1836645522-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f0a7a688391c4fbe9e1088c5047f00d6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ab2e9f5-54fd-4cab-9405-ed65e2aaba64", "external-id": "nsx-vlan-transportzone-222", "segmentation_id": 222, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51efbac7-c4", "ovs_interfaceid": "51efbac7-c441-4555-9a3c-d42167d75fcc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 794.480177] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a7:a6:af', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6ab2e9f5-54fd-4cab-9405-ed65e2aaba64', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '51efbac7-c441-4555-9a3c-d42167d75fcc', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 794.489245] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Creating folder: Project (f0a7a688391c4fbe9e1088c5047f00d6). Parent ref: group-v281134. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 794.489520] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-618eced6-4c5a-4d32-a138-5cc30cb7720b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.500430] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Created folder: Project (f0a7a688391c4fbe9e1088c5047f00d6) in parent group-v281134. [ 794.500627] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Creating folder: Instances. Parent ref: group-v281161. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 794.500862] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0a596fb4-a247-4896-9f12-383507ef4a48 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.509810] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Created folder: Instances in parent group-v281161. [ 794.510061] env[62460]: DEBUG oslo.service.loopingcall [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 794.510267] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 794.510472] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9fd4d24d-8f06-455a-9b75-395dbadc2a53 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.529384] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 794.529384] env[62460]: value = "task-1313492" [ 794.529384] env[62460]: _type = "Task" [ 794.529384] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.536914] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313492, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.572749] env[62460]: DEBUG nova.compute.manager [req-fb270304-ab0c-47b5-90ca-f6504e74cfca req-bff35c95-2700-4d37-a6e1-eb17a28c8ab3 service nova] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Received event network-changed-51efbac7-c441-4555-9a3c-d42167d75fcc {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 794.572947] env[62460]: DEBUG nova.compute.manager [req-fb270304-ab0c-47b5-90ca-f6504e74cfca req-bff35c95-2700-4d37-a6e1-eb17a28c8ab3 service nova] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Refreshing instance network info cache due to event network-changed-51efbac7-c441-4555-9a3c-d42167d75fcc. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 794.573174] env[62460]: DEBUG oslo_concurrency.lockutils [req-fb270304-ab0c-47b5-90ca-f6504e74cfca req-bff35c95-2700-4d37-a6e1-eb17a28c8ab3 service nova] Acquiring lock "refresh_cache-1f318a64-2c38-470b-8fae-4ba4543a5681" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.573320] env[62460]: DEBUG oslo_concurrency.lockutils [req-fb270304-ab0c-47b5-90ca-f6504e74cfca req-bff35c95-2700-4d37-a6e1-eb17a28c8ab3 service nova] Acquired lock "refresh_cache-1f318a64-2c38-470b-8fae-4ba4543a5681" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.573544] env[62460]: DEBUG nova.network.neutron [req-fb270304-ab0c-47b5-90ca-f6504e74cfca req-bff35c95-2700-4d37-a6e1-eb17a28c8ab3 service nova] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Refreshing network info cache for port 51efbac7-c441-4555-9a3c-d42167d75fcc {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 794.649553] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a515bdf6-c358-49d9-9607-0fb06eafdb17 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Lock "51718896-f5bf-43a9-9396-1ac768737ba2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 109.977s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.671364] env[62460]: INFO nova.compute.manager [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Rebuilding instance [ 794.719235] env[62460]: DEBUG nova.compute.manager [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 794.719921] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9f407ce-f7d2-4048-8b90-248f15f3a949 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.858307] env[62460]: DEBUG nova.network.neutron [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 794.998143] env[62460]: DEBUG nova.network.neutron [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.040530] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313492, 'name': CreateVM_Task} progress is 25%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.073219] env[62460]: DEBUG nova.network.neutron [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Successfully updated port: ba38cbb0-62cf-47cb-a238-87ebc03de250 {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 795.152017] env[62460]: DEBUG nova.compute.manager [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 795.232477] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 795.232786] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1b0d9bf5-bfce-4023-a8e0-28d83e12b2e0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.239718] env[62460]: DEBUG oslo_vmware.api [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Waiting for the task: (returnval){ [ 795.239718] env[62460]: value = "task-1313493" [ 795.239718] env[62460]: _type = "Task" [ 795.239718] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.248672] env[62460]: DEBUG oslo_vmware.api [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Task: {'id': task-1313493, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.356989] env[62460]: DEBUG nova.network.neutron [req-fb270304-ab0c-47b5-90ca-f6504e74cfca req-bff35c95-2700-4d37-a6e1-eb17a28c8ab3 service nova] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Updated VIF entry in instance network info cache for port 51efbac7-c441-4555-9a3c-d42167d75fcc. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 795.357545] env[62460]: DEBUG nova.network.neutron [req-fb270304-ab0c-47b5-90ca-f6504e74cfca req-bff35c95-2700-4d37-a6e1-eb17a28c8ab3 service nova] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Updating instance_info_cache with network_info: [{"id": "51efbac7-c441-4555-9a3c-d42167d75fcc", "address": "fa:16:3e:a7:a6:af", "network": {"id": "6d99b7f4-65ae-4a8f-aa6e-bce5b845dbe1", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1836645522-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f0a7a688391c4fbe9e1088c5047f00d6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ab2e9f5-54fd-4cab-9405-ed65e2aaba64", "external-id": "nsx-vlan-transportzone-222", "segmentation_id": 222, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51efbac7-c4", "ovs_interfaceid": "51efbac7-c441-4555-9a3c-d42167d75fcc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.501166] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Releasing lock "refresh_cache-588e5489-2f61-473a-90e9-883eb8b163c8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 795.503592] env[62460]: DEBUG nova.compute.manager [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 795.503592] env[62460]: DEBUG nova.compute.manager [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 795.503592] env[62460]: DEBUG nova.network.neutron [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 795.519502] env[62460]: DEBUG nova.network.neutron [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 795.543327] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313492, 'name': CreateVM_Task, 'duration_secs': 0.787308} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.543494] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 795.554078] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 795.554078] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.554078] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 795.554078] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2f0ee10d-9e8a-4c6e-b0df-12236ce8067f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.556524] env[62460]: DEBUG oslo_vmware.api [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Waiting for the task: (returnval){ [ 795.556524] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]528c3a05-f7da-deaa-ddec-d93dc7d4fdf1" [ 795.556524] env[62460]: _type = "Task" [ 795.556524] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.565928] env[62460]: DEBUG oslo_vmware.api [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]528c3a05-f7da-deaa-ddec-d93dc7d4fdf1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.579526] env[62460]: DEBUG oslo_concurrency.lockutils [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "refresh_cache-7e6ff902-4a04-43d5-9014-38c4ec88efc4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 795.579681] env[62460]: DEBUG oslo_concurrency.lockutils [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquired lock "refresh_cache-7e6ff902-4a04-43d5-9014-38c4ec88efc4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.579837] env[62460]: DEBUG nova.network.neutron [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 795.663604] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca470f9f-9521-44a9-b436-21fe04ee992d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.675709] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e4c9976-83e2-4f17-ae59-ddc4b2c5e5cd {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.680573] env[62460]: DEBUG oslo_concurrency.lockutils [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.713803] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e56bd00-4dbf-4cbc-ac65-a01c66e1a82f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.721902] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50a32a01-ae79-49ad-bb33-59162520792d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.735974] env[62460]: DEBUG nova.compute.provider_tree [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 795.749357] env[62460]: DEBUG oslo_vmware.api [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Task: {'id': task-1313493, 'name': PowerOffVM_Task, 'duration_secs': 0.147267} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.749676] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 795.750009] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 795.750830] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbe73c38-e3ed-4257-8b6b-42e506372473 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.760039] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 795.760455] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-06f41f82-3fc4-43af-8916-905442e1e4a9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.785174] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 795.785606] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Deleting contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 795.785843] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Deleting the datastore file [datastore1] 51718896-f5bf-43a9-9396-1ac768737ba2 {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 795.786121] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-90730182-a14f-4268-b13d-328ea9adbe5f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.792706] env[62460]: DEBUG oslo_vmware.api [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Waiting for the task: (returnval){ [ 795.792706] env[62460]: value = "task-1313496" [ 795.792706] env[62460]: _type = "Task" [ 795.792706] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.800652] env[62460]: DEBUG oslo_vmware.api [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Task: {'id': task-1313496, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.862412] env[62460]: DEBUG oslo_concurrency.lockutils [req-fb270304-ab0c-47b5-90ca-f6504e74cfca req-bff35c95-2700-4d37-a6e1-eb17a28c8ab3 service nova] Releasing lock "refresh_cache-1f318a64-2c38-470b-8fae-4ba4543a5681" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.027639] env[62460]: DEBUG nova.network.neutron [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 796.067112] env[62460]: DEBUG oslo_vmware.api [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]528c3a05-f7da-deaa-ddec-d93dc7d4fdf1, 'name': SearchDatastore_Task, 'duration_secs': 0.018344} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.067531] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.067830] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 796.068512] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.068512] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.068512] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 796.068850] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0fc87496-b462-4aaa-8f77-4a03c431402a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.077195] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 796.077381] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 796.078866] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-684d72f2-b73a-453b-9927-87139f2f91a5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.085454] env[62460]: DEBUG oslo_vmware.api [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Waiting for the task: (returnval){ [ 796.085454] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52a0aeaf-ad31-5bf5-da04-d0aa1d747045" [ 796.085454] env[62460]: _type = "Task" [ 796.085454] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.098101] env[62460]: DEBUG oslo_vmware.api [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52a0aeaf-ad31-5bf5-da04-d0aa1d747045, 'name': SearchDatastore_Task, 'duration_secs': 0.008382} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.099066] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-15c99afe-121b-449a-8154-a5cf4f465432 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.104353] env[62460]: DEBUG oslo_vmware.api [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Waiting for the task: (returnval){ [ 796.104353] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]5201aa3e-1097-5080-75fe-22843f5bf9ac" [ 796.104353] env[62460]: _type = "Task" [ 796.104353] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.112541] env[62460]: DEBUG oslo_vmware.api [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5201aa3e-1097-5080-75fe-22843f5bf9ac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.217697] env[62460]: DEBUG nova.network.neutron [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 796.245216] env[62460]: DEBUG nova.scheduler.client.report [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 796.304129] env[62460]: DEBUG oslo_vmware.api [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Task: {'id': task-1313496, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.08498} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.306293] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 796.306491] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Deleted contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 796.306673] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 796.398759] env[62460]: DEBUG nova.network.neutron [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Updating instance_info_cache with network_info: [{"id": "ba38cbb0-62cf-47cb-a238-87ebc03de250", "address": "fa:16:3e:b4:93:18", "network": {"id": "d3136e32-ad55-4b73-835c-8fa5f0480767", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1511414968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53fa06008e9f43488362895e7a143700", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "054fcd1e-638e-425a-a1de-78cb188ae026", "external-id": "nsx-vlan-transportzone-658", "segmentation_id": 658, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba38cbb0-62", "ovs_interfaceid": "ba38cbb0-62cf-47cb-a238-87ebc03de250", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 796.531603] env[62460]: INFO nova.compute.manager [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] [instance: 588e5489-2f61-473a-90e9-883eb8b163c8] Took 1.03 seconds to deallocate network for instance. [ 796.600137] env[62460]: DEBUG nova.compute.manager [req-f3c076fa-2acb-447e-a75c-81c0b8c315a5 req-3a513745-7ea5-4cac-b1d2-1ba6d36a9721 service nova] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Received event network-vif-plugged-ba38cbb0-62cf-47cb-a238-87ebc03de250 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 796.600137] env[62460]: DEBUG oslo_concurrency.lockutils [req-f3c076fa-2acb-447e-a75c-81c0b8c315a5 req-3a513745-7ea5-4cac-b1d2-1ba6d36a9721 service nova] Acquiring lock "7e6ff902-4a04-43d5-9014-38c4ec88efc4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.600137] env[62460]: DEBUG oslo_concurrency.lockutils [req-f3c076fa-2acb-447e-a75c-81c0b8c315a5 req-3a513745-7ea5-4cac-b1d2-1ba6d36a9721 service nova] Lock "7e6ff902-4a04-43d5-9014-38c4ec88efc4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.600137] env[62460]: DEBUG oslo_concurrency.lockutils [req-f3c076fa-2acb-447e-a75c-81c0b8c315a5 req-3a513745-7ea5-4cac-b1d2-1ba6d36a9721 service nova] Lock "7e6ff902-4a04-43d5-9014-38c4ec88efc4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.600137] env[62460]: DEBUG nova.compute.manager [req-f3c076fa-2acb-447e-a75c-81c0b8c315a5 req-3a513745-7ea5-4cac-b1d2-1ba6d36a9721 service nova] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] No waiting events found dispatching network-vif-plugged-ba38cbb0-62cf-47cb-a238-87ebc03de250 {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 796.600735] env[62460]: WARNING nova.compute.manager [req-f3c076fa-2acb-447e-a75c-81c0b8c315a5 req-3a513745-7ea5-4cac-b1d2-1ba6d36a9721 service nova] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Received unexpected event network-vif-plugged-ba38cbb0-62cf-47cb-a238-87ebc03de250 for instance with vm_state building and task_state spawning. [ 796.600735] env[62460]: DEBUG nova.compute.manager [req-f3c076fa-2acb-447e-a75c-81c0b8c315a5 req-3a513745-7ea5-4cac-b1d2-1ba6d36a9721 service nova] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Received event network-changed-ba38cbb0-62cf-47cb-a238-87ebc03de250 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 796.600735] env[62460]: DEBUG nova.compute.manager [req-f3c076fa-2acb-447e-a75c-81c0b8c315a5 req-3a513745-7ea5-4cac-b1d2-1ba6d36a9721 service nova] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Refreshing instance network info cache due to event network-changed-ba38cbb0-62cf-47cb-a238-87ebc03de250. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 796.600735] env[62460]: DEBUG oslo_concurrency.lockutils [req-f3c076fa-2acb-447e-a75c-81c0b8c315a5 req-3a513745-7ea5-4cac-b1d2-1ba6d36a9721 service nova] Acquiring lock "refresh_cache-7e6ff902-4a04-43d5-9014-38c4ec88efc4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.617335] env[62460]: DEBUG oslo_vmware.api [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5201aa3e-1097-5080-75fe-22843f5bf9ac, 'name': SearchDatastore_Task, 'duration_secs': 0.007545} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.617335] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.617779] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 1f318a64-2c38-470b-8fae-4ba4543a5681/1f318a64-2c38-470b-8fae-4ba4543a5681.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 796.620023] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ce80a3d5-d37b-4a73-a7d3-6ec5963f34f8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.628023] env[62460]: DEBUG oslo_vmware.api [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Waiting for the task: (returnval){ [ 796.628023] env[62460]: value = "task-1313497" [ 796.628023] env[62460]: _type = "Task" [ 796.628023] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.635570] env[62460]: DEBUG oslo_vmware.api [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Task: {'id': task-1313497, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.755027] env[62460]: DEBUG oslo_concurrency.lockutils [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.433s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.755027] env[62460]: DEBUG nova.compute.manager [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 796.758243] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 17.415s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.901812] env[62460]: DEBUG oslo_concurrency.lockutils [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Releasing lock "refresh_cache-7e6ff902-4a04-43d5-9014-38c4ec88efc4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.902205] env[62460]: DEBUG nova.compute.manager [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Instance network_info: |[{"id": "ba38cbb0-62cf-47cb-a238-87ebc03de250", "address": "fa:16:3e:b4:93:18", "network": {"id": "d3136e32-ad55-4b73-835c-8fa5f0480767", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1511414968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53fa06008e9f43488362895e7a143700", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "054fcd1e-638e-425a-a1de-78cb188ae026", "external-id": "nsx-vlan-transportzone-658", "segmentation_id": 658, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba38cbb0-62", "ovs_interfaceid": "ba38cbb0-62cf-47cb-a238-87ebc03de250", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 796.902663] env[62460]: DEBUG oslo_concurrency.lockutils [req-f3c076fa-2acb-447e-a75c-81c0b8c315a5 req-3a513745-7ea5-4cac-b1d2-1ba6d36a9721 service nova] Acquired lock "refresh_cache-7e6ff902-4a04-43d5-9014-38c4ec88efc4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.902875] env[62460]: DEBUG nova.network.neutron [req-f3c076fa-2acb-447e-a75c-81c0b8c315a5 req-3a513745-7ea5-4cac-b1d2-1ba6d36a9721 service nova] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Refreshing network info cache for port ba38cbb0-62cf-47cb-a238-87ebc03de250 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 796.904415] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b4:93:18', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '054fcd1e-638e-425a-a1de-78cb188ae026', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ba38cbb0-62cf-47cb-a238-87ebc03de250', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 796.912602] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Creating folder: Project (53fa06008e9f43488362895e7a143700). Parent ref: group-v281134. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 796.916023] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-59aef67a-433d-4835-8e75-9a0cce8428d3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.926917] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Created folder: Project (53fa06008e9f43488362895e7a143700) in parent group-v281134. [ 796.927188] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Creating folder: Instances. Parent ref: group-v281164. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 796.927533] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ce4aa194-0dc8-46e4-a3f4-9e37309d8dd5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.936564] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Created folder: Instances in parent group-v281164. [ 796.936564] env[62460]: DEBUG oslo.service.loopingcall [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 796.936564] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 796.936768] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1aa2a7bf-cf92-4878-850b-967e3517606a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.959822] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 796.959822] env[62460]: value = "task-1313500" [ 796.959822] env[62460]: _type = "Task" [ 796.959822] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.968944] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313500, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.137112] env[62460]: DEBUG oslo_vmware.api [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Task: {'id': task-1313497, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.494472} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.137112] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 1f318a64-2c38-470b-8fae-4ba4543a5681/1f318a64-2c38-470b-8fae-4ba4543a5681.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 797.137425] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 797.137584] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c084eaa0-9bdd-4b53-8b64-dd69df33f81b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.143855] env[62460]: DEBUG oslo_vmware.api [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Waiting for the task: (returnval){ [ 797.143855] env[62460]: value = "task-1313501" [ 797.143855] env[62460]: _type = "Task" [ 797.143855] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.151650] env[62460]: DEBUG oslo_vmware.api [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Task: {'id': task-1313501, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.156148] env[62460]: DEBUG oslo_concurrency.lockutils [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Acquiring lock "49477d35-92ea-4f9d-8333-fc23144b7dfc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.156390] env[62460]: DEBUG oslo_concurrency.lockutils [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Lock "49477d35-92ea-4f9d-8333-fc23144b7dfc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.199228] env[62460]: DEBUG nova.network.neutron [req-f3c076fa-2acb-447e-a75c-81c0b8c315a5 req-3a513745-7ea5-4cac-b1d2-1ba6d36a9721 service nova] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Updated VIF entry in instance network info cache for port ba38cbb0-62cf-47cb-a238-87ebc03de250. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 797.199599] env[62460]: DEBUG nova.network.neutron [req-f3c076fa-2acb-447e-a75c-81c0b8c315a5 req-3a513745-7ea5-4cac-b1d2-1ba6d36a9721 service nova] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Updating instance_info_cache with network_info: [{"id": "ba38cbb0-62cf-47cb-a238-87ebc03de250", "address": "fa:16:3e:b4:93:18", "network": {"id": "d3136e32-ad55-4b73-835c-8fa5f0480767", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1511414968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53fa06008e9f43488362895e7a143700", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "054fcd1e-638e-425a-a1de-78cb188ae026", "external-id": "nsx-vlan-transportzone-658", "segmentation_id": 658, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba38cbb0-62", "ovs_interfaceid": "ba38cbb0-62cf-47cb-a238-87ebc03de250", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.260384] env[62460]: DEBUG nova.compute.utils [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 797.261852] env[62460]: DEBUG nova.compute.manager [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 797.262039] env[62460]: DEBUG nova.network.neutron [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 797.307309] env[62460]: DEBUG nova.policy [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '631d6b4e25944471964d830066931857', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '309249b168af45e9869d6122f5703a3d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 797.348734] env[62460]: DEBUG nova.virt.hardware [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 797.348987] env[62460]: DEBUG nova.virt.hardware [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 797.349164] env[62460]: DEBUG nova.virt.hardware [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 797.349350] env[62460]: DEBUG nova.virt.hardware [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 797.349568] env[62460]: DEBUG nova.virt.hardware [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 797.349653] env[62460]: DEBUG nova.virt.hardware [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 797.349859] env[62460]: DEBUG nova.virt.hardware [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 797.350047] env[62460]: DEBUG nova.virt.hardware [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 797.350237] env[62460]: DEBUG nova.virt.hardware [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 797.350408] env[62460]: DEBUG nova.virt.hardware [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 797.350584] env[62460]: DEBUG nova.virt.hardware [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 797.351464] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a0386f9-f335-4c27-a1a1-e4e336785f9b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.359611] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ab17747-0322-4781-aab7-deeef298aa0a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.372789] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Instance VIF info [] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 797.378171] env[62460]: DEBUG oslo.service.loopingcall [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 797.378414] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 797.378924] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ba74d4ac-5031-41e9-8a18-a9f8ebf16064 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.394640] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 797.394640] env[62460]: value = "task-1313502" [ 797.394640] env[62460]: _type = "Task" [ 797.394640] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.405653] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313502, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.469918] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313500, 'name': CreateVM_Task, 'duration_secs': 0.370946} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.470101] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 797.470762] env[62460]: DEBUG oslo_concurrency.lockutils [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 797.470931] env[62460]: DEBUG oslo_concurrency.lockutils [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.471267] env[62460]: DEBUG oslo_concurrency.lockutils [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 797.471523] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7742a39c-37d1-466d-a8bb-ae31f78aafb6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.476100] env[62460]: DEBUG oslo_vmware.api [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Waiting for the task: (returnval){ [ 797.476100] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52c2fdd0-8e54-9ecb-6336-6657a3d9a906" [ 797.476100] env[62460]: _type = "Task" [ 797.476100] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.483843] env[62460]: DEBUG oslo_vmware.api [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52c2fdd0-8e54-9ecb-6336-6657a3d9a906, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.570117] env[62460]: INFO nova.scheduler.client.report [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Deleted allocations for instance 588e5489-2f61-473a-90e9-883eb8b163c8 [ 797.611158] env[62460]: DEBUG nova.network.neutron [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Successfully created port: 213b9859-4858-4268-a921-2a6a5d9e1f8a {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 797.654109] env[62460]: DEBUG oslo_vmware.api [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Task: {'id': task-1313501, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084902} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.654503] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 797.655369] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7d379ce-ea02-4789-ba9e-6f7bca37746c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.678581] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Reconfiguring VM instance instance-0000003a to attach disk [datastore1] 1f318a64-2c38-470b-8fae-4ba4543a5681/1f318a64-2c38-470b-8fae-4ba4543a5681.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 797.679712] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-965c6a38-a39d-4982-ba42-ec3510f82a12 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.700809] env[62460]: DEBUG oslo_vmware.api [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Waiting for the task: (returnval){ [ 797.700809] env[62460]: value = "task-1313504" [ 797.700809] env[62460]: _type = "Task" [ 797.700809] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.705906] env[62460]: DEBUG oslo_concurrency.lockutils [req-f3c076fa-2acb-447e-a75c-81c0b8c315a5 req-3a513745-7ea5-4cac-b1d2-1ba6d36a9721 service nova] Releasing lock "refresh_cache-7e6ff902-4a04-43d5-9014-38c4ec88efc4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 797.711638] env[62460]: DEBUG oslo_vmware.api [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Task: {'id': task-1313504, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.769733] env[62460]: DEBUG nova.compute.manager [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 797.795732] env[62460]: WARNING nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 4ed90d16-81a6-4dbd-8936-0e137151171f is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 797.795884] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 6c58f9fe-fb6d-4012-8b69-39e4134996b2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 797.796021] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance c7d51b39-b449-4af3-a4d0-c746983ded3e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 797.796150] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 51718896-f5bf-43a9-9396-1ac768737ba2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 797.796262] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 1f318a64-2c38-470b-8fae-4ba4543a5681 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 797.796377] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 7e6ff902-4a04-43d5-9014-38c4ec88efc4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 797.796488] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance db09d1f5-88cc-4dc7-9a7b-5d53d09567fa actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 797.904747] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313502, 'name': CreateVM_Task, 'duration_secs': 0.310374} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.904929] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 797.905389] env[62460]: DEBUG oslo_concurrency.lockutils [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 797.905573] env[62460]: DEBUG oslo_concurrency.lockutils [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.905904] env[62460]: DEBUG oslo_concurrency.lockutils [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 797.906183] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9c7e1ad2-2aa2-4f78-8d82-46c8cfc008fe {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.910670] env[62460]: DEBUG oslo_vmware.api [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Waiting for the task: (returnval){ [ 797.910670] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]5250e896-8490-d399-24a0-d422042e4d1e" [ 797.910670] env[62460]: _type = "Task" [ 797.910670] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.920788] env[62460]: DEBUG oslo_vmware.api [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5250e896-8490-d399-24a0-d422042e4d1e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.987067] env[62460]: DEBUG oslo_vmware.api [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52c2fdd0-8e54-9ecb-6336-6657a3d9a906, 'name': SearchDatastore_Task, 'duration_secs': 0.014554} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.987347] env[62460]: DEBUG oslo_concurrency.lockutils [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 797.987762] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 797.987823] env[62460]: DEBUG oslo_concurrency.lockutils [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 797.987946] env[62460]: DEBUG oslo_concurrency.lockutils [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.988144] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 797.988434] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c4f57cfe-5282-4d9b-b0d4-19c9ec1237cf {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.997287] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 797.997561] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 797.998351] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-17bad6ff-ce5c-4a32-8cb4-2fa355b4f704 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.003994] env[62460]: DEBUG oslo_vmware.api [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Waiting for the task: (returnval){ [ 798.003994] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52f8b29b-3acd-34ad-c2cd-c48e0d0d5c88" [ 798.003994] env[62460]: _type = "Task" [ 798.003994] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.012360] env[62460]: DEBUG oslo_vmware.api [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52f8b29b-3acd-34ad-c2cd-c48e0d0d5c88, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.082610] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3f6c6afe-58e1-4039-876d-4640814a7847 tempest-VolumesAdminNegativeTest-1686493033 tempest-VolumesAdminNegativeTest-1686493033-project-member] Lock "588e5489-2f61-473a-90e9-883eb8b163c8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 119.409s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.213283] env[62460]: DEBUG oslo_vmware.api [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Task: {'id': task-1313504, 'name': ReconfigVM_Task, 'duration_secs': 0.25652} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.213376] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Reconfigured VM instance instance-0000003a to attach disk [datastore1] 1f318a64-2c38-470b-8fae-4ba4543a5681/1f318a64-2c38-470b-8fae-4ba4543a5681.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 798.215302] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a88ed037-65e8-4a67-b78d-c7a3a7cd4852 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.219633] env[62460]: DEBUG oslo_vmware.api [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Waiting for the task: (returnval){ [ 798.219633] env[62460]: value = "task-1313505" [ 798.219633] env[62460]: _type = "Task" [ 798.219633] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.227886] env[62460]: DEBUG oslo_vmware.api [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Task: {'id': task-1313505, 'name': Rename_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.299781] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 72e91f8e-0619-464c-b9bc-d6a14be42cb8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 798.423885] env[62460]: DEBUG oslo_vmware.api [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5250e896-8490-d399-24a0-d422042e4d1e, 'name': SearchDatastore_Task, 'duration_secs': 0.009539} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.424208] env[62460]: DEBUG oslo_concurrency.lockutils [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.424471] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 798.424992] env[62460]: DEBUG oslo_concurrency.lockutils [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.424992] env[62460]: DEBUG oslo_concurrency.lockutils [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.425141] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 798.425393] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6b8b96e4-04be-40d4-8b90-ccc20d8bd881 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.433015] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 798.433467] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 798.433952] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-75309cdc-c29e-428c-b3d4-5910139e0ed1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.441292] env[62460]: DEBUG oslo_vmware.api [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Waiting for the task: (returnval){ [ 798.441292] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52b36cd3-3609-3d3d-2905-04350401f95c" [ 798.441292] env[62460]: _type = "Task" [ 798.441292] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.447074] env[62460]: DEBUG oslo_vmware.api [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52b36cd3-3609-3d3d-2905-04350401f95c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.514331] env[62460]: DEBUG oslo_vmware.api [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52f8b29b-3acd-34ad-c2cd-c48e0d0d5c88, 'name': SearchDatastore_Task, 'duration_secs': 0.009176} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.515147] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c56021df-ebe9-4bbb-ba6c-44510f78fa27 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.521075] env[62460]: DEBUG oslo_vmware.api [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Waiting for the task: (returnval){ [ 798.521075] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52ea466a-c98c-6687-0a6d-f6df74cbf660" [ 798.521075] env[62460]: _type = "Task" [ 798.521075] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.527605] env[62460]: DEBUG oslo_vmware.api [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52ea466a-c98c-6687-0a6d-f6df74cbf660, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.584817] env[62460]: DEBUG nova.compute.manager [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 798.729988] env[62460]: DEBUG oslo_vmware.api [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Task: {'id': task-1313505, 'name': Rename_Task, 'duration_secs': 0.130308} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.732702] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 798.732702] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-92b2e57e-3026-49c9-87b4-0dfe42b33b79 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.737990] env[62460]: DEBUG oslo_vmware.api [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Waiting for the task: (returnval){ [ 798.737990] env[62460]: value = "task-1313506" [ 798.737990] env[62460]: _type = "Task" [ 798.737990] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.747326] env[62460]: DEBUG oslo_vmware.api [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Task: {'id': task-1313506, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.783059] env[62460]: DEBUG nova.compute.manager [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 798.803306] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 3ada3516-3147-4566-a46a-1cb29cf880d0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 798.812089] env[62460]: DEBUG nova.virt.hardware [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 798.812353] env[62460]: DEBUG nova.virt.hardware [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 798.812536] env[62460]: DEBUG nova.virt.hardware [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 798.812718] env[62460]: DEBUG nova.virt.hardware [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 798.812868] env[62460]: DEBUG nova.virt.hardware [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 798.813032] env[62460]: DEBUG nova.virt.hardware [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 798.813300] env[62460]: DEBUG nova.virt.hardware [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 798.813438] env[62460]: DEBUG nova.virt.hardware [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 798.813613] env[62460]: DEBUG nova.virt.hardware [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 798.813816] env[62460]: DEBUG nova.virt.hardware [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 798.814031] env[62460]: DEBUG nova.virt.hardware [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 798.814920] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39cae0fa-ac36-4b45-8739-938031368fbd {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.825677] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cd250fc-cce5-40b3-9524-fab081f40375 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.949952] env[62460]: DEBUG oslo_vmware.api [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52b36cd3-3609-3d3d-2905-04350401f95c, 'name': SearchDatastore_Task, 'duration_secs': 0.008775} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.950728] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-81a8834e-c981-40df-aa19-a37456af877b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.957017] env[62460]: DEBUG oslo_vmware.api [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Waiting for the task: (returnval){ [ 798.957017] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]529d7c9e-f413-464c-57df-83bab01485d2" [ 798.957017] env[62460]: _type = "Task" [ 798.957017] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.964689] env[62460]: DEBUG oslo_vmware.api [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]529d7c9e-f413-464c-57df-83bab01485d2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.030420] env[62460]: DEBUG oslo_vmware.api [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52ea466a-c98c-6687-0a6d-f6df74cbf660, 'name': SearchDatastore_Task, 'duration_secs': 0.008626} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.030529] env[62460]: DEBUG oslo_concurrency.lockutils [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.030783] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 7e6ff902-4a04-43d5-9014-38c4ec88efc4/7e6ff902-4a04-43d5-9014-38c4ec88efc4.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 799.031053] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-51c7cd4e-04dc-4b1b-b11c-555519fdea89 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.037851] env[62460]: DEBUG oslo_vmware.api [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Waiting for the task: (returnval){ [ 799.037851] env[62460]: value = "task-1313507" [ 799.037851] env[62460]: _type = "Task" [ 799.037851] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.046160] env[62460]: DEBUG oslo_vmware.api [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313507, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.111454] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 799.260977] env[62460]: DEBUG oslo_vmware.api [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Task: {'id': task-1313506, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.307182] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance fde12685-d3b9-46a0-8931-25b904d4f21e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 799.395537] env[62460]: DEBUG nova.compute.manager [req-16a762fc-957f-49f8-b999-65de5325f5d7 req-9cb78ab1-ea9c-40ac-8132-55c09f612ae9 service nova] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Received event network-vif-plugged-213b9859-4858-4268-a921-2a6a5d9e1f8a {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 799.395782] env[62460]: DEBUG oslo_concurrency.lockutils [req-16a762fc-957f-49f8-b999-65de5325f5d7 req-9cb78ab1-ea9c-40ac-8132-55c09f612ae9 service nova] Acquiring lock "db09d1f5-88cc-4dc7-9a7b-5d53d09567fa-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 799.396075] env[62460]: DEBUG oslo_concurrency.lockutils [req-16a762fc-957f-49f8-b999-65de5325f5d7 req-9cb78ab1-ea9c-40ac-8132-55c09f612ae9 service nova] Lock "db09d1f5-88cc-4dc7-9a7b-5d53d09567fa-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 799.396471] env[62460]: DEBUG oslo_concurrency.lockutils [req-16a762fc-957f-49f8-b999-65de5325f5d7 req-9cb78ab1-ea9c-40ac-8132-55c09f612ae9 service nova] Lock "db09d1f5-88cc-4dc7-9a7b-5d53d09567fa-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.396664] env[62460]: DEBUG nova.compute.manager [req-16a762fc-957f-49f8-b999-65de5325f5d7 req-9cb78ab1-ea9c-40ac-8132-55c09f612ae9 service nova] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] No waiting events found dispatching network-vif-plugged-213b9859-4858-4268-a921-2a6a5d9e1f8a {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 799.396898] env[62460]: WARNING nova.compute.manager [req-16a762fc-957f-49f8-b999-65de5325f5d7 req-9cb78ab1-ea9c-40ac-8132-55c09f612ae9 service nova] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Received unexpected event network-vif-plugged-213b9859-4858-4268-a921-2a6a5d9e1f8a for instance with vm_state building and task_state spawning. [ 799.477078] env[62460]: DEBUG oslo_vmware.api [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]529d7c9e-f413-464c-57df-83bab01485d2, 'name': SearchDatastore_Task, 'duration_secs': 0.009124} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.477378] env[62460]: DEBUG oslo_concurrency.lockutils [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.477708] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] 51718896-f5bf-43a9-9396-1ac768737ba2/51718896-f5bf-43a9-9396-1ac768737ba2.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 799.477917] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-349a62ec-6bb7-4427-812c-358f1c483c0b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.486886] env[62460]: DEBUG oslo_vmware.api [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Waiting for the task: (returnval){ [ 799.486886] env[62460]: value = "task-1313508" [ 799.486886] env[62460]: _type = "Task" [ 799.486886] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.497143] env[62460]: DEBUG oslo_vmware.api [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Task: {'id': task-1313508, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.551296] env[62460]: DEBUG oslo_vmware.api [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313507, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.501589} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.553721] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 7e6ff902-4a04-43d5-9014-38c4ec88efc4/7e6ff902-4a04-43d5-9014-38c4ec88efc4.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 799.554878] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 799.556550] env[62460]: DEBUG nova.network.neutron [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Successfully updated port: 213b9859-4858-4268-a921-2a6a5d9e1f8a {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 799.558557] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f4b39c22-259f-4495-8eaf-fed4da76cdbe {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.572862] env[62460]: DEBUG oslo_vmware.api [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Waiting for the task: (returnval){ [ 799.572862] env[62460]: value = "task-1313509" [ 799.572862] env[62460]: _type = "Task" [ 799.572862] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.582701] env[62460]: DEBUG oslo_vmware.api [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313509, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.758014] env[62460]: DEBUG oslo_vmware.api [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Task: {'id': task-1313506, 'name': PowerOnVM_Task, 'duration_secs': 0.549417} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.758014] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 799.758014] env[62460]: INFO nova.compute.manager [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Took 7.76 seconds to spawn the instance on the hypervisor. [ 799.758014] env[62460]: DEBUG nova.compute.manager [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 799.758585] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df379281-36dc-4445-8a94-2bee81b73c44 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.813021] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 13890351-6091-4b4f-8484-1cdd0c8523b0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 799.997425] env[62460]: DEBUG oslo_vmware.api [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Task: {'id': task-1313508, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.498612} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.997855] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] 51718896-f5bf-43a9-9396-1ac768737ba2/51718896-f5bf-43a9-9396-1ac768737ba2.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 799.998254] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 799.998603] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-19fe771b-2968-473a-b866-0114a0d59e57 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.004746] env[62460]: DEBUG oslo_vmware.api [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Waiting for the task: (returnval){ [ 800.004746] env[62460]: value = "task-1313511" [ 800.004746] env[62460]: _type = "Task" [ 800.004746] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.013340] env[62460]: DEBUG oslo_vmware.api [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Task: {'id': task-1313511, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.072121] env[62460]: DEBUG oslo_concurrency.lockutils [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Acquiring lock "refresh_cache-db09d1f5-88cc-4dc7-9a7b-5d53d09567fa" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.072121] env[62460]: DEBUG oslo_concurrency.lockutils [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Acquired lock "refresh_cache-db09d1f5-88cc-4dc7-9a7b-5d53d09567fa" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.072121] env[62460]: DEBUG nova.network.neutron [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 800.085784] env[62460]: DEBUG oslo_vmware.api [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313509, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061553} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.085784] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 800.085784] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-287c38f3-fd38-4249-939b-e61b8bb8395a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.106976] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Reconfiguring VM instance instance-0000003b to attach disk [datastore1] 7e6ff902-4a04-43d5-9014-38c4ec88efc4/7e6ff902-4a04-43d5-9014-38c4ec88efc4.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 800.108106] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-72396cdb-e7af-45bb-a068-53c8d378c33d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.128901] env[62460]: DEBUG oslo_vmware.api [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Waiting for the task: (returnval){ [ 800.128901] env[62460]: value = "task-1313512" [ 800.128901] env[62460]: _type = "Task" [ 800.128901] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.138365] env[62460]: DEBUG oslo_vmware.api [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313512, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.286027] env[62460]: INFO nova.compute.manager [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Took 30.70 seconds to build instance. [ 800.316481] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance b76f37a0-91d0-4a01-9d95-9c6586081175 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 800.518037] env[62460]: DEBUG oslo_vmware.api [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Task: {'id': task-1313511, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067081} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.518326] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 800.520443] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e24cafdb-8fe0-454f-beca-52dcd968e2e7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.541058] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Reconfiguring VM instance instance-00000039 to attach disk [datastore2] 51718896-f5bf-43a9-9396-1ac768737ba2/51718896-f5bf-43a9-9396-1ac768737ba2.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 800.541058] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b95da787-6c33-4123-917a-4bf77ed0d8ef {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.559868] env[62460]: DEBUG oslo_vmware.api [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Waiting for the task: (returnval){ [ 800.559868] env[62460]: value = "task-1313513" [ 800.559868] env[62460]: _type = "Task" [ 800.559868] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.568284] env[62460]: DEBUG oslo_vmware.api [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Task: {'id': task-1313513, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.635055] env[62460]: DEBUG nova.network.neutron [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 800.643036] env[62460]: DEBUG oslo_vmware.api [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313512, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.786858] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2b9424c1-bf71-414e-a46b-4179ac276112 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Lock "1f318a64-2c38-470b-8fae-4ba4543a5681" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 115.100s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.797795] env[62460]: DEBUG nova.network.neutron [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Updating instance_info_cache with network_info: [{"id": "213b9859-4858-4268-a921-2a6a5d9e1f8a", "address": "fa:16:3e:ea:2a:4e", "network": {"id": "c81ba1c6-6f9f-4817-83fa-42f156766144", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1930612639-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "309249b168af45e9869d6122f5703a3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db1f7867-8524-469c-ab47-d2c9e2751d98", "external-id": "nsx-vlan-transportzone-130", "segmentation_id": 130, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap213b9859-48", "ovs_interfaceid": "213b9859-4858-4268-a921-2a6a5d9e1f8a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 800.820572] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 5b6e8205-003e-49c0-a73d-be2e032a8272 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 801.069610] env[62460]: DEBUG oslo_vmware.api [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Task: {'id': task-1313513, 'name': ReconfigVM_Task, 'duration_secs': 0.286242} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.070076] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Reconfigured VM instance instance-00000039 to attach disk [datastore2] 51718896-f5bf-43a9-9396-1ac768737ba2/51718896-f5bf-43a9-9396-1ac768737ba2.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 801.070798] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c0237330-f46c-4a1b-913f-be09283878b7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.077122] env[62460]: DEBUG oslo_vmware.api [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Waiting for the task: (returnval){ [ 801.077122] env[62460]: value = "task-1313514" [ 801.077122] env[62460]: _type = "Task" [ 801.077122] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.085921] env[62460]: DEBUG oslo_vmware.api [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Task: {'id': task-1313514, 'name': Rename_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.139476] env[62460]: DEBUG oslo_vmware.api [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313512, 'name': ReconfigVM_Task, 'duration_secs': 0.748289} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.139975] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Reconfigured VM instance instance-0000003b to attach disk [datastore1] 7e6ff902-4a04-43d5-9014-38c4ec88efc4/7e6ff902-4a04-43d5-9014-38c4ec88efc4.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 801.140795] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8ab3cabb-eb6f-49a6-a07a-c54874d4efa5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.147191] env[62460]: DEBUG oslo_vmware.api [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Waiting for the task: (returnval){ [ 801.147191] env[62460]: value = "task-1313515" [ 801.147191] env[62460]: _type = "Task" [ 801.147191] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.156318] env[62460]: DEBUG oslo_vmware.api [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313515, 'name': Rename_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.291785] env[62460]: DEBUG nova.compute.manager [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 801.304016] env[62460]: DEBUG oslo_concurrency.lockutils [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Releasing lock "refresh_cache-db09d1f5-88cc-4dc7-9a7b-5d53d09567fa" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.304016] env[62460]: DEBUG nova.compute.manager [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Instance network_info: |[{"id": "213b9859-4858-4268-a921-2a6a5d9e1f8a", "address": "fa:16:3e:ea:2a:4e", "network": {"id": "c81ba1c6-6f9f-4817-83fa-42f156766144", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1930612639-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "309249b168af45e9869d6122f5703a3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db1f7867-8524-469c-ab47-d2c9e2751d98", "external-id": "nsx-vlan-transportzone-130", "segmentation_id": 130, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap213b9859-48", "ovs_interfaceid": "213b9859-4858-4268-a921-2a6a5d9e1f8a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 801.304247] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ea:2a:4e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'db1f7867-8524-469c-ab47-d2c9e2751d98', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '213b9859-4858-4268-a921-2a6a5d9e1f8a', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 801.309941] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Creating folder: Project (309249b168af45e9869d6122f5703a3d). Parent ref: group-v281134. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 801.310372] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ad127b89-c650-4695-b6a1-1a0f5efe3a0e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.321317] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Created folder: Project (309249b168af45e9869d6122f5703a3d) in parent group-v281134. [ 801.321664] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Creating folder: Instances. Parent ref: group-v281169. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 801.322410] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 801.323740] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-57f450f5-1092-417f-bd68-d5deeefcfbc2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.333286] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Created folder: Instances in parent group-v281169. [ 801.333684] env[62460]: DEBUG oslo.service.loopingcall [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 801.333999] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 801.334319] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c1cd79ed-6969-4095-bdba-67b303731f82 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.354637] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 801.354637] env[62460]: value = "task-1313518" [ 801.354637] env[62460]: _type = "Task" [ 801.354637] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.362788] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313518, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.587945] env[62460]: DEBUG oslo_vmware.api [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Task: {'id': task-1313514, 'name': Rename_Task, 'duration_secs': 0.127607} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.587945] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 801.588119] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0a7ffb58-ac97-4759-a005-e09d9c081faa {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.592912] env[62460]: DEBUG nova.compute.manager [req-a9405636-0198-4b03-8de6-1bd0f1329edb req-e883c524-e5c4-45ae-b577-b4565b727003 service nova] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Received event network-changed-213b9859-4858-4268-a921-2a6a5d9e1f8a {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 801.593115] env[62460]: DEBUG nova.compute.manager [req-a9405636-0198-4b03-8de6-1bd0f1329edb req-e883c524-e5c4-45ae-b577-b4565b727003 service nova] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Refreshing instance network info cache due to event network-changed-213b9859-4858-4268-a921-2a6a5d9e1f8a. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 801.593328] env[62460]: DEBUG oslo_concurrency.lockutils [req-a9405636-0198-4b03-8de6-1bd0f1329edb req-e883c524-e5c4-45ae-b577-b4565b727003 service nova] Acquiring lock "refresh_cache-db09d1f5-88cc-4dc7-9a7b-5d53d09567fa" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.593472] env[62460]: DEBUG oslo_concurrency.lockutils [req-a9405636-0198-4b03-8de6-1bd0f1329edb req-e883c524-e5c4-45ae-b577-b4565b727003 service nova] Acquired lock "refresh_cache-db09d1f5-88cc-4dc7-9a7b-5d53d09567fa" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.593627] env[62460]: DEBUG nova.network.neutron [req-a9405636-0198-4b03-8de6-1bd0f1329edb req-e883c524-e5c4-45ae-b577-b4565b727003 service nova] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Refreshing network info cache for port 213b9859-4858-4268-a921-2a6a5d9e1f8a {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 801.596106] env[62460]: DEBUG oslo_vmware.api [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Waiting for the task: (returnval){ [ 801.596106] env[62460]: value = "task-1313519" [ 801.596106] env[62460]: _type = "Task" [ 801.596106] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.604243] env[62460]: DEBUG oslo_vmware.api [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Task: {'id': task-1313519, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.657763] env[62460]: DEBUG oslo_vmware.api [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313515, 'name': Rename_Task, 'duration_secs': 0.315946} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.657763] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 801.657763] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4d93c56b-0aa4-478d-ba19-025760175604 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.664784] env[62460]: DEBUG oslo_vmware.api [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Waiting for the task: (returnval){ [ 801.664784] env[62460]: value = "task-1313520" [ 801.664784] env[62460]: _type = "Task" [ 801.664784] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.673094] env[62460]: DEBUG oslo_vmware.api [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313520, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.822058] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 801.828146] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 0269dc64-d2b1-43c5-bdf7-11d97e534819 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 801.868971] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313518, 'name': CreateVM_Task, 'duration_secs': 0.486256} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.869237] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 801.869952] env[62460]: DEBUG oslo_concurrency.lockutils [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.870201] env[62460]: DEBUG oslo_concurrency.lockutils [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.870604] env[62460]: DEBUG oslo_concurrency.lockutils [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 801.871015] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a347622e-2bfa-4151-b376-e520cf46191c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.876120] env[62460]: DEBUG oslo_vmware.api [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Waiting for the task: (returnval){ [ 801.876120] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52fcdce8-9f30-7f34-0184-36043d3bdd01" [ 801.876120] env[62460]: _type = "Task" [ 801.876120] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.885837] env[62460]: DEBUG oslo_vmware.api [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52fcdce8-9f30-7f34-0184-36043d3bdd01, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.022944] env[62460]: DEBUG nova.compute.manager [req-fbbe8d73-541b-4f01-807a-f5869e1d8a6e req-69e86e3a-6141-472b-b8d7-381c4503d88d service nova] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Received event network-changed-51efbac7-c441-4555-9a3c-d42167d75fcc {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 802.024214] env[62460]: DEBUG nova.compute.manager [req-fbbe8d73-541b-4f01-807a-f5869e1d8a6e req-69e86e3a-6141-472b-b8d7-381c4503d88d service nova] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Refreshing instance network info cache due to event network-changed-51efbac7-c441-4555-9a3c-d42167d75fcc. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 802.024571] env[62460]: DEBUG oslo_concurrency.lockutils [req-fbbe8d73-541b-4f01-807a-f5869e1d8a6e req-69e86e3a-6141-472b-b8d7-381c4503d88d service nova] Acquiring lock "refresh_cache-1f318a64-2c38-470b-8fae-4ba4543a5681" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.026829] env[62460]: DEBUG oslo_concurrency.lockutils [req-fbbe8d73-541b-4f01-807a-f5869e1d8a6e req-69e86e3a-6141-472b-b8d7-381c4503d88d service nova] Acquired lock "refresh_cache-1f318a64-2c38-470b-8fae-4ba4543a5681" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.027039] env[62460]: DEBUG nova.network.neutron [req-fbbe8d73-541b-4f01-807a-f5869e1d8a6e req-69e86e3a-6141-472b-b8d7-381c4503d88d service nova] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Refreshing network info cache for port 51efbac7-c441-4555-9a3c-d42167d75fcc {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 802.108472] env[62460]: DEBUG oslo_vmware.api [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Task: {'id': task-1313519, 'name': PowerOnVM_Task, 'duration_secs': 0.416062} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.108758] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 802.108975] env[62460]: DEBUG nova.compute.manager [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 802.109769] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e499ccf-3b2d-4e21-8970-1e4e591d5645 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.176460] env[62460]: DEBUG oslo_vmware.api [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313520, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.331523] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 9d1a5830-f3c0-4d18-9338-16f7b6962c6a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 802.388134] env[62460]: DEBUG oslo_vmware.api [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52fcdce8-9f30-7f34-0184-36043d3bdd01, 'name': SearchDatastore_Task, 'duration_secs': 0.009902} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.388488] env[62460]: DEBUG oslo_concurrency.lockutils [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.388735] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 802.389016] env[62460]: DEBUG oslo_concurrency.lockutils [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.389176] env[62460]: DEBUG oslo_concurrency.lockutils [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.389367] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 802.389632] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3e66543d-a20d-4653-9e5c-d7af3533ebc8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.398224] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 802.398409] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 802.399144] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2e5da9bb-a576-4de6-80d3-36397febb35c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.405252] env[62460]: DEBUG oslo_vmware.api [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Waiting for the task: (returnval){ [ 802.405252] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]525ba124-81e8-5127-5bad-3e1ae61b2827" [ 802.405252] env[62460]: _type = "Task" [ 802.405252] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.412382] env[62460]: DEBUG oslo_vmware.api [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]525ba124-81e8-5127-5bad-3e1ae61b2827, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.449857] env[62460]: DEBUG nova.network.neutron [req-a9405636-0198-4b03-8de6-1bd0f1329edb req-e883c524-e5c4-45ae-b577-b4565b727003 service nova] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Updated VIF entry in instance network info cache for port 213b9859-4858-4268-a921-2a6a5d9e1f8a. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 802.450250] env[62460]: DEBUG nova.network.neutron [req-a9405636-0198-4b03-8de6-1bd0f1329edb req-e883c524-e5c4-45ae-b577-b4565b727003 service nova] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Updating instance_info_cache with network_info: [{"id": "213b9859-4858-4268-a921-2a6a5d9e1f8a", "address": "fa:16:3e:ea:2a:4e", "network": {"id": "c81ba1c6-6f9f-4817-83fa-42f156766144", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1930612639-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "309249b168af45e9869d6122f5703a3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db1f7867-8524-469c-ab47-d2c9e2751d98", "external-id": "nsx-vlan-transportzone-130", "segmentation_id": 130, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap213b9859-48", "ovs_interfaceid": "213b9859-4858-4268-a921-2a6a5d9e1f8a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 802.625428] env[62460]: DEBUG oslo_concurrency.lockutils [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.676258] env[62460]: DEBUG oslo_vmware.api [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313520, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.829364] env[62460]: DEBUG nova.network.neutron [req-fbbe8d73-541b-4f01-807a-f5869e1d8a6e req-69e86e3a-6141-472b-b8d7-381c4503d88d service nova] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Updated VIF entry in instance network info cache for port 51efbac7-c441-4555-9a3c-d42167d75fcc. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 802.829364] env[62460]: DEBUG nova.network.neutron [req-fbbe8d73-541b-4f01-807a-f5869e1d8a6e req-69e86e3a-6141-472b-b8d7-381c4503d88d service nova] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Updating instance_info_cache with network_info: [{"id": "51efbac7-c441-4555-9a3c-d42167d75fcc", "address": "fa:16:3e:a7:a6:af", "network": {"id": "6d99b7f4-65ae-4a8f-aa6e-bce5b845dbe1", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1836645522-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.200", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f0a7a688391c4fbe9e1088c5047f00d6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ab2e9f5-54fd-4cab-9405-ed65e2aaba64", "external-id": "nsx-vlan-transportzone-222", "segmentation_id": 222, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51efbac7-c4", "ovs_interfaceid": "51efbac7-c441-4555-9a3c-d42167d75fcc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 802.838172] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance d7d180f7-e1a8-46c8-ba8e-ca50dac474cc has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 802.917123] env[62460]: DEBUG oslo_vmware.api [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]525ba124-81e8-5127-5bad-3e1ae61b2827, 'name': SearchDatastore_Task, 'duration_secs': 0.008967} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.918099] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-83d9f646-3a3a-4fb9-90ff-4f93f5825ce9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.923688] env[62460]: DEBUG oslo_vmware.api [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Waiting for the task: (returnval){ [ 802.923688] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]522a02a6-11f1-65f6-71a7-080eb1855077" [ 802.923688] env[62460]: _type = "Task" [ 802.923688] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.931138] env[62460]: DEBUG oslo_vmware.api [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]522a02a6-11f1-65f6-71a7-080eb1855077, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.952956] env[62460]: DEBUG oslo_concurrency.lockutils [req-a9405636-0198-4b03-8de6-1bd0f1329edb req-e883c524-e5c4-45ae-b577-b4565b727003 service nova] Releasing lock "refresh_cache-db09d1f5-88cc-4dc7-9a7b-5d53d09567fa" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.181609] env[62460]: DEBUG oslo_vmware.api [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313520, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.337507] env[62460]: DEBUG oslo_concurrency.lockutils [req-fbbe8d73-541b-4f01-807a-f5869e1d8a6e req-69e86e3a-6141-472b-b8d7-381c4503d88d service nova] Releasing lock "refresh_cache-1f318a64-2c38-470b-8fae-4ba4543a5681" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.339697] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 16d40829-ba6e-4193-98b5-fff94c066bcb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 803.435220] env[62460]: DEBUG oslo_vmware.api [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]522a02a6-11f1-65f6-71a7-080eb1855077, 'name': SearchDatastore_Task, 'duration_secs': 0.009452} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.435220] env[62460]: DEBUG oslo_concurrency.lockutils [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.435398] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] db09d1f5-88cc-4dc7-9a7b-5d53d09567fa/db09d1f5-88cc-4dc7-9a7b-5d53d09567fa.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 803.435654] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-96de74b8-564f-4e6a-be29-b58763382274 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.442055] env[62460]: DEBUG oslo_vmware.api [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Waiting for the task: (returnval){ [ 803.442055] env[62460]: value = "task-1313522" [ 803.442055] env[62460]: _type = "Task" [ 803.442055] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.449484] env[62460]: DEBUG oslo_vmware.api [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Task: {'id': task-1313522, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.596845] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7b25a346-a87d-486c-86ee-f9dc9c62cc1e tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Acquiring lock "51718896-f5bf-43a9-9396-1ac768737ba2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 803.597186] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7b25a346-a87d-486c-86ee-f9dc9c62cc1e tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Lock "51718896-f5bf-43a9-9396-1ac768737ba2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 803.597407] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7b25a346-a87d-486c-86ee-f9dc9c62cc1e tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Acquiring lock "51718896-f5bf-43a9-9396-1ac768737ba2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 803.597592] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7b25a346-a87d-486c-86ee-f9dc9c62cc1e tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Lock "51718896-f5bf-43a9-9396-1ac768737ba2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 803.597763] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7b25a346-a87d-486c-86ee-f9dc9c62cc1e tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Lock "51718896-f5bf-43a9-9396-1ac768737ba2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.599927] env[62460]: INFO nova.compute.manager [None req-7b25a346-a87d-486c-86ee-f9dc9c62cc1e tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Terminating instance [ 803.601666] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7b25a346-a87d-486c-86ee-f9dc9c62cc1e tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Acquiring lock "refresh_cache-51718896-f5bf-43a9-9396-1ac768737ba2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 803.601825] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7b25a346-a87d-486c-86ee-f9dc9c62cc1e tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Acquired lock "refresh_cache-51718896-f5bf-43a9-9396-1ac768737ba2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.601990] env[62460]: DEBUG nova.network.neutron [None req-7b25a346-a87d-486c-86ee-f9dc9c62cc1e tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 803.679384] env[62460]: DEBUG oslo_vmware.api [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313520, 'name': PowerOnVM_Task} progress is 90%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.843778] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance a02b4be1-91b0-4254-8d60-654885e24f6b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 803.959845] env[62460]: DEBUG oslo_vmware.api [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Task: {'id': task-1313522, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.121265] env[62460]: DEBUG nova.network.neutron [None req-7b25a346-a87d-486c-86ee-f9dc9c62cc1e tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 804.176822] env[62460]: DEBUG nova.network.neutron [None req-7b25a346-a87d-486c-86ee-f9dc9c62cc1e tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.186473] env[62460]: DEBUG oslo_vmware.api [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313520, 'name': PowerOnVM_Task, 'duration_secs': 2.071969} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.188738] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7b25a346-a87d-486c-86ee-f9dc9c62cc1e tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Releasing lock "refresh_cache-51718896-f5bf-43a9-9396-1ac768737ba2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.189194] env[62460]: DEBUG nova.compute.manager [None req-7b25a346-a87d-486c-86ee-f9dc9c62cc1e tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 804.189408] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7b25a346-a87d-486c-86ee-f9dc9c62cc1e tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 804.189695] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 804.189888] env[62460]: INFO nova.compute.manager [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Took 9.78 seconds to spawn the instance on the hypervisor. [ 804.190079] env[62460]: DEBUG nova.compute.manager [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 804.191144] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8000f9b-d7ce-4f21-9bb2-94b5104af696 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.194370] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-887b0e6c-2805-4db3-9448-7a53d25a7b23 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.210395] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b25a346-a87d-486c-86ee-f9dc9c62cc1e tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 804.211039] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-afd6b437-a872-4768-95b4-34b5946a3f87 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.218031] env[62460]: DEBUG oslo_vmware.api [None req-7b25a346-a87d-486c-86ee-f9dc9c62cc1e tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Waiting for the task: (returnval){ [ 804.218031] env[62460]: value = "task-1313524" [ 804.218031] env[62460]: _type = "Task" [ 804.218031] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.232037] env[62460]: DEBUG oslo_vmware.api [None req-7b25a346-a87d-486c-86ee-f9dc9c62cc1e tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Task: {'id': task-1313524, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.346385] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 49477d35-92ea-4f9d-8333-fc23144b7dfc has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 804.346650] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=62460) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 804.346813] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1664MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=62460) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 804.453380] env[62460]: DEBUG oslo_vmware.api [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Task: {'id': task-1313522, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.569891} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.453652] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] db09d1f5-88cc-4dc7-9a7b-5d53d09567fa/db09d1f5-88cc-4dc7-9a7b-5d53d09567fa.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 804.453872] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 804.454138] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b66720fe-06d1-4621-803e-09ff8388e5ba {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.464113] env[62460]: DEBUG oslo_vmware.api [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Waiting for the task: (returnval){ [ 804.464113] env[62460]: value = "task-1313525" [ 804.464113] env[62460]: _type = "Task" [ 804.464113] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.476592] env[62460]: DEBUG oslo_vmware.api [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Task: {'id': task-1313525, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.600463] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64a57034-db48-4e53-9e8d-fe87b544be19 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.607857] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-220130af-73d4-4072-898f-d1fbb3135429 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.640291] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21b4287e-5549-4b4f-a050-952e6b301a9a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.649441] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-846313cc-6795-4dcb-9772-bfafced3797c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.665545] env[62460]: DEBUG nova.compute.provider_tree [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 804.722402] env[62460]: INFO nova.compute.manager [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Took 32.98 seconds to build instance. [ 804.728989] env[62460]: DEBUG oslo_vmware.api [None req-7b25a346-a87d-486c-86ee-f9dc9c62cc1e tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Task: {'id': task-1313524, 'name': PowerOffVM_Task, 'duration_secs': 0.257383} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.729533] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b25a346-a87d-486c-86ee-f9dc9c62cc1e tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 804.729751] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7b25a346-a87d-486c-86ee-f9dc9c62cc1e tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 804.730082] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e6673d79-a45e-46af-81ee-ad47279f2249 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.753847] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7b25a346-a87d-486c-86ee-f9dc9c62cc1e tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 804.754092] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7b25a346-a87d-486c-86ee-f9dc9c62cc1e tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Deleting contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 804.754550] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b25a346-a87d-486c-86ee-f9dc9c62cc1e tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Deleting the datastore file [datastore2] 51718896-f5bf-43a9-9396-1ac768737ba2 {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 804.754828] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8316eab3-bd92-44e3-b201-038c22d3b57f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.762347] env[62460]: DEBUG oslo_vmware.api [None req-7b25a346-a87d-486c-86ee-f9dc9c62cc1e tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Waiting for the task: (returnval){ [ 804.762347] env[62460]: value = "task-1313527" [ 804.762347] env[62460]: _type = "Task" [ 804.762347] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.770163] env[62460]: DEBUG oslo_vmware.api [None req-7b25a346-a87d-486c-86ee-f9dc9c62cc1e tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Task: {'id': task-1313527, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.979871] env[62460]: DEBUG oslo_vmware.api [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Task: {'id': task-1313525, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.2974} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.980181] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 804.980952] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42910c37-b608-4321-81a9-f332bc33f7d7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.009182] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Reconfiguring VM instance instance-0000003c to attach disk [datastore2] db09d1f5-88cc-4dc7-9a7b-5d53d09567fa/db09d1f5-88cc-4dc7-9a7b-5d53d09567fa.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 805.011722] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e50a787d-6b81-46f1-931d-268660a551b8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.033437] env[62460]: DEBUG oslo_vmware.api [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Waiting for the task: (returnval){ [ 805.033437] env[62460]: value = "task-1313528" [ 805.033437] env[62460]: _type = "Task" [ 805.033437] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.043296] env[62460]: DEBUG oslo_vmware.api [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Task: {'id': task-1313528, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.170025] env[62460]: DEBUG nova.scheduler.client.report [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 805.225433] env[62460]: DEBUG oslo_concurrency.lockutils [None req-402d31e3-1eda-4de9-8e63-d23c5b17d7bf tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "7e6ff902-4a04-43d5-9014-38c4ec88efc4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 115.234s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.272293] env[62460]: DEBUG oslo_vmware.api [None req-7b25a346-a87d-486c-86ee-f9dc9c62cc1e tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Task: {'id': task-1313527, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.42981} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.272693] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b25a346-a87d-486c-86ee-f9dc9c62cc1e tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 805.272905] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7b25a346-a87d-486c-86ee-f9dc9c62cc1e tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Deleted contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 805.273096] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7b25a346-a87d-486c-86ee-f9dc9c62cc1e tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 805.273268] env[62460]: INFO nova.compute.manager [None req-7b25a346-a87d-486c-86ee-f9dc9c62cc1e tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Took 1.08 seconds to destroy the instance on the hypervisor. [ 805.273503] env[62460]: DEBUG oslo.service.loopingcall [None req-7b25a346-a87d-486c-86ee-f9dc9c62cc1e tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 805.273684] env[62460]: DEBUG nova.compute.manager [-] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 805.273773] env[62460]: DEBUG nova.network.neutron [-] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 805.292281] env[62460]: DEBUG nova.network.neutron [-] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 805.543499] env[62460]: DEBUG oslo_vmware.api [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Task: {'id': task-1313528, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.675658] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62460) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 805.676010] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 8.918s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.677966] env[62460]: DEBUG oslo_concurrency.lockutils [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 26.023s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.688018] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 805.688018] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Cleaning up deleted instances {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 805.728169] env[62460]: DEBUG nova.compute.manager [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 805.780618] env[62460]: DEBUG nova.compute.manager [req-3abf21a2-1609-4ef7-9467-18c78b7e23ed req-a6029bd5-228f-4f0a-b3e8-0ad2e119f98d service nova] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Received event network-changed-ba38cbb0-62cf-47cb-a238-87ebc03de250 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 805.780618] env[62460]: DEBUG nova.compute.manager [req-3abf21a2-1609-4ef7-9467-18c78b7e23ed req-a6029bd5-228f-4f0a-b3e8-0ad2e119f98d service nova] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Refreshing instance network info cache due to event network-changed-ba38cbb0-62cf-47cb-a238-87ebc03de250. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 805.780871] env[62460]: DEBUG oslo_concurrency.lockutils [req-3abf21a2-1609-4ef7-9467-18c78b7e23ed req-a6029bd5-228f-4f0a-b3e8-0ad2e119f98d service nova] Acquiring lock "refresh_cache-7e6ff902-4a04-43d5-9014-38c4ec88efc4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.780938] env[62460]: DEBUG oslo_concurrency.lockutils [req-3abf21a2-1609-4ef7-9467-18c78b7e23ed req-a6029bd5-228f-4f0a-b3e8-0ad2e119f98d service nova] Acquired lock "refresh_cache-7e6ff902-4a04-43d5-9014-38c4ec88efc4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.781101] env[62460]: DEBUG nova.network.neutron [req-3abf21a2-1609-4ef7-9467-18c78b7e23ed req-a6029bd5-228f-4f0a-b3e8-0ad2e119f98d service nova] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Refreshing network info cache for port ba38cbb0-62cf-47cb-a238-87ebc03de250 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 805.794045] env[62460]: DEBUG nova.network.neutron [-] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.044539] env[62460]: DEBUG oslo_vmware.api [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Task: {'id': task-1313528, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.193325] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] There are 4 instances to clean {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 806.194486] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 5214f4a1-3e28-41bf-88d9-161511385e1b] Instance has had 0 of 5 cleanup attempts {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 806.255894] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.297245] env[62460]: INFO nova.compute.manager [-] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Took 1.02 seconds to deallocate network for instance. [ 806.532415] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b9aab6a-877f-4e43-a81f-dffbd7a83f6a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.546034] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a13bbab9-7e9c-4e79-8446-1a26032cc58b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.554704] env[62460]: DEBUG oslo_vmware.api [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Task: {'id': task-1313528, 'name': ReconfigVM_Task, 'duration_secs': 1.034477} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.555385] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Reconfigured VM instance instance-0000003c to attach disk [datastore2] db09d1f5-88cc-4dc7-9a7b-5d53d09567fa/db09d1f5-88cc-4dc7-9a7b-5d53d09567fa.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 806.556033] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b958674a-c5c7-492f-b767-dfb81b31fd04 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.590680] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac6cad95-2732-4694-af0f-db79ee348e9f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.599391] env[62460]: DEBUG oslo_vmware.api [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Waiting for the task: (returnval){ [ 806.599391] env[62460]: value = "task-1313530" [ 806.599391] env[62460]: _type = "Task" [ 806.599391] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.604768] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22fcc3b5-decb-4425-8aed-0821545f9d37 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.612030] env[62460]: DEBUG oslo_vmware.api [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Task: {'id': task-1313530, 'name': Rename_Task} progress is 6%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.622107] env[62460]: DEBUG nova.compute.provider_tree [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 806.645452] env[62460]: DEBUG nova.network.neutron [req-3abf21a2-1609-4ef7-9467-18c78b7e23ed req-a6029bd5-228f-4f0a-b3e8-0ad2e119f98d service nova] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Updated VIF entry in instance network info cache for port ba38cbb0-62cf-47cb-a238-87ebc03de250. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 806.645890] env[62460]: DEBUG nova.network.neutron [req-3abf21a2-1609-4ef7-9467-18c78b7e23ed req-a6029bd5-228f-4f0a-b3e8-0ad2e119f98d service nova] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Updating instance_info_cache with network_info: [{"id": "ba38cbb0-62cf-47cb-a238-87ebc03de250", "address": "fa:16:3e:b4:93:18", "network": {"id": "d3136e32-ad55-4b73-835c-8fa5f0480767", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1511414968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53fa06008e9f43488362895e7a143700", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "054fcd1e-638e-425a-a1de-78cb188ae026", "external-id": "nsx-vlan-transportzone-658", "segmentation_id": 658, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba38cbb0-62", "ovs_interfaceid": "ba38cbb0-62cf-47cb-a238-87ebc03de250", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.703637] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 3b71c366-cbd9-4b98-aa0d-c55b56e69231] Instance has had 0 of 5 cleanup attempts {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 806.806646] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7b25a346-a87d-486c-86ee-f9dc9c62cc1e tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.108924] env[62460]: DEBUG oslo_vmware.api [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Task: {'id': task-1313530, 'name': Rename_Task, 'duration_secs': 0.179281} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.109206] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 807.109473] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3a31dfec-c61e-4e8d-9345-e5c3c3954d53 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.117628] env[62460]: DEBUG oslo_vmware.api [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Waiting for the task: (returnval){ [ 807.117628] env[62460]: value = "task-1313531" [ 807.117628] env[62460]: _type = "Task" [ 807.117628] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.127918] env[62460]: DEBUG nova.scheduler.client.report [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 807.131900] env[62460]: DEBUG oslo_vmware.api [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Task: {'id': task-1313531, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.150560] env[62460]: DEBUG oslo_concurrency.lockutils [req-3abf21a2-1609-4ef7-9467-18c78b7e23ed req-a6029bd5-228f-4f0a-b3e8-0ad2e119f98d service nova] Releasing lock "refresh_cache-7e6ff902-4a04-43d5-9014-38c4ec88efc4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.208372] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 00544d2a-1a15-4347-abe4-3641b5bf0cfe] Instance has had 0 of 5 cleanup attempts {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 807.629683] env[62460]: DEBUG oslo_vmware.api [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Task: {'id': task-1313531, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.633608] env[62460]: DEBUG oslo_concurrency.lockutils [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.956s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.634237] env[62460]: ERROR nova.compute.manager [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d957de37-a110-4cb6-b807-78c130f4de9e, please check neutron logs for more information. [ 807.634237] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] Traceback (most recent call last): [ 807.634237] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 807.634237] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] self.driver.spawn(context, instance, image_meta, [ 807.634237] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 807.634237] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 807.634237] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 807.634237] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] vm_ref = self.build_virtual_machine(instance, [ 807.634237] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 807.634237] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] vif_infos = vmwarevif.get_vif_info(self._session, [ 807.634237] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 807.634520] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] for vif in network_info: [ 807.634520] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 807.634520] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] return self._sync_wrapper(fn, *args, **kwargs) [ 807.634520] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 807.634520] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] self.wait() [ 807.634520] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 807.634520] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] self[:] = self._gt.wait() [ 807.634520] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 807.634520] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] return self._exit_event.wait() [ 807.634520] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 807.634520] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] result = hub.switch() [ 807.634520] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 807.634520] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] return self.greenlet.switch() [ 807.634842] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 807.634842] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] result = function(*args, **kwargs) [ 807.634842] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 807.634842] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] return func(*args, **kwargs) [ 807.634842] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 807.634842] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] raise e [ 807.634842] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 807.634842] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] nwinfo = self.network_api.allocate_for_instance( [ 807.634842] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 807.634842] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] created_port_ids = self._update_ports_for_instance( [ 807.634842] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 807.634842] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] with excutils.save_and_reraise_exception(): [ 807.634842] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 807.635145] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] self.force_reraise() [ 807.635145] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 807.635145] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] raise self.value [ 807.635145] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 807.635145] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] updated_port = self._update_port( [ 807.635145] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 807.635145] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] _ensure_no_port_binding_failure(port) [ 807.635145] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 807.635145] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] raise exception.PortBindingFailed(port_id=port['id']) [ 807.635145] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] nova.exception.PortBindingFailed: Binding failed for port d957de37-a110-4cb6-b807-78c130f4de9e, please check neutron logs for more information. [ 807.635145] env[62460]: ERROR nova.compute.manager [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] [ 807.635520] env[62460]: DEBUG nova.compute.utils [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] Binding failed for port d957de37-a110-4cb6-b807-78c130f4de9e, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 807.636327] env[62460]: DEBUG oslo_concurrency.lockutils [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.638s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.637749] env[62460]: INFO nova.compute.claims [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 807.641651] env[62460]: DEBUG nova.compute.manager [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] Build of instance 6c58f9fe-fb6d-4012-8b69-39e4134996b2 was re-scheduled: Binding failed for port d957de37-a110-4cb6-b807-78c130f4de9e, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 807.641651] env[62460]: DEBUG nova.compute.manager [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 807.641651] env[62460]: DEBUG oslo_concurrency.lockutils [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] Acquiring lock "refresh_cache-6c58f9fe-fb6d-4012-8b69-39e4134996b2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.641651] env[62460]: DEBUG oslo_concurrency.lockutils [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] Acquired lock "refresh_cache-6c58f9fe-fb6d-4012-8b69-39e4134996b2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.641846] env[62460]: DEBUG nova.network.neutron [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 807.711986] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 44a2ac83-3fff-4958-a1d3-b3884b42c211] Instance has had 0 of 5 cleanup attempts {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 808.027494] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquiring lock "896f8ff4-e45f-4403-a727-03ee25e58609" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.027731] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lock "896f8ff4-e45f-4403-a727-03ee25e58609" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.129308] env[62460]: DEBUG oslo_vmware.api [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Task: {'id': task-1313531, 'name': PowerOnVM_Task, 'duration_secs': 0.890809} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.129786] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 808.129863] env[62460]: INFO nova.compute.manager [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Took 9.35 seconds to spawn the instance on the hypervisor. [ 808.130151] env[62460]: DEBUG nova.compute.manager [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 808.131041] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c3cb78c-cada-4561-88c3-f8c28d734dd2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.160975] env[62460]: DEBUG nova.network.neutron [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 808.215625] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 808.215625] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Cleaning up deleted instances with incomplete migration {{(pid=62460) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 808.250758] env[62460]: DEBUG nova.network.neutron [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.651445] env[62460]: INFO nova.compute.manager [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Took 32.09 seconds to build instance. [ 808.717738] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 808.753407] env[62460]: DEBUG oslo_concurrency.lockutils [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] Releasing lock "refresh_cache-6c58f9fe-fb6d-4012-8b69-39e4134996b2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.754054] env[62460]: DEBUG nova.compute.manager [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 808.754054] env[62460]: DEBUG nova.compute.manager [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 808.754054] env[62460]: DEBUG nova.network.neutron [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 808.777480] env[62460]: DEBUG nova.network.neutron [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 808.921928] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b63a18da-b75f-4ab3-9e9f-d3c51839bb40 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.930185] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f0c0a14-6324-4298-a14b-5b4ddf0cb2a3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.960832] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c540bff1-1a92-458d-8a01-efa4694b7786 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.968637] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e7a73a3-10dc-4231-a9cb-b109435883eb {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.982242] env[62460]: DEBUG nova.compute.provider_tree [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 809.156323] env[62460]: DEBUG oslo_concurrency.lockutils [None req-89eab27e-f6fc-45f0-b6f4-bb4c2b7e579d tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Lock "db09d1f5-88cc-4dc7-9a7b-5d53d09567fa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 97.769s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.283816] env[62460]: DEBUG nova.network.neutron [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.486406] env[62460]: DEBUG nova.scheduler.client.report [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 809.662140] env[62460]: DEBUG nova.compute.manager [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 809.667120] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9c83d044-4a1e-4d35-a1ca-cefdd159ac66 tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Acquiring lock "db09d1f5-88cc-4dc7-9a7b-5d53d09567fa" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.667450] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9c83d044-4a1e-4d35-a1ca-cefdd159ac66 tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Lock "db09d1f5-88cc-4dc7-9a7b-5d53d09567fa" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.667711] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9c83d044-4a1e-4d35-a1ca-cefdd159ac66 tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Acquiring lock "db09d1f5-88cc-4dc7-9a7b-5d53d09567fa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.667958] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9c83d044-4a1e-4d35-a1ca-cefdd159ac66 tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Lock "db09d1f5-88cc-4dc7-9a7b-5d53d09567fa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.668204] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9c83d044-4a1e-4d35-a1ca-cefdd159ac66 tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Lock "db09d1f5-88cc-4dc7-9a7b-5d53d09567fa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.670440] env[62460]: INFO nova.compute.manager [None req-9c83d044-4a1e-4d35-a1ca-cefdd159ac66 tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Terminating instance [ 809.672319] env[62460]: DEBUG nova.compute.manager [None req-9c83d044-4a1e-4d35-a1ca-cefdd159ac66 tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 809.672569] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-9c83d044-4a1e-4d35-a1ca-cefdd159ac66 tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 809.673489] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a424ecca-2075-4f82-8d32-24067e635054 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.681362] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c83d044-4a1e-4d35-a1ca-cefdd159ac66 tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 809.681594] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b56f7d24-d31c-41a9-a314-d22ccf657c58 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.687359] env[62460]: DEBUG oslo_vmware.api [None req-9c83d044-4a1e-4d35-a1ca-cefdd159ac66 tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Waiting for the task: (returnval){ [ 809.687359] env[62460]: value = "task-1313532" [ 809.687359] env[62460]: _type = "Task" [ 809.687359] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.695416] env[62460]: DEBUG oslo_vmware.api [None req-9c83d044-4a1e-4d35-a1ca-cefdd159ac66 tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Task: {'id': task-1313532, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.787053] env[62460]: INFO nova.compute.manager [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] [instance: 6c58f9fe-fb6d-4012-8b69-39e4134996b2] Took 1.03 seconds to deallocate network for instance. [ 809.895053] env[62460]: DEBUG oslo_concurrency.lockutils [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Acquiring lock "f1921907-f69c-49a3-896c-a4b12b526ffa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.895331] env[62460]: DEBUG oslo_concurrency.lockutils [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Lock "f1921907-f69c-49a3-896c-a4b12b526ffa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.992675] env[62460]: DEBUG oslo_concurrency.lockutils [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.356s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.993214] env[62460]: DEBUG nova.compute.manager [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 809.995963] env[62460]: DEBUG oslo_concurrency.lockutils [None req-cb805937-3829-46aa-b450-89395080ba46 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.513s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.996169] env[62460]: DEBUG oslo_concurrency.lockutils [None req-cb805937-3829-46aa-b450-89395080ba46 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.998226] env[62460]: DEBUG oslo_concurrency.lockutils [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.195s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.999924] env[62460]: INFO nova.compute.claims [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 810.018964] env[62460]: INFO nova.scheduler.client.report [None req-cb805937-3829-46aa-b450-89395080ba46 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Deleted allocations for instance 4ed90d16-81a6-4dbd-8936-0e137151171f [ 810.186688] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.197354] env[62460]: DEBUG oslo_vmware.api [None req-9c83d044-4a1e-4d35-a1ca-cefdd159ac66 tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Task: {'id': task-1313532, 'name': PowerOffVM_Task, 'duration_secs': 0.182171} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.197602] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c83d044-4a1e-4d35-a1ca-cefdd159ac66 tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 810.197782] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-9c83d044-4a1e-4d35-a1ca-cefdd159ac66 tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 810.198066] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9d57be18-8a22-463d-991f-c6df5219bcd8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.257221] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-9c83d044-4a1e-4d35-a1ca-cefdd159ac66 tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 810.257453] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-9c83d044-4a1e-4d35-a1ca-cefdd159ac66 tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Deleting contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 810.257635] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c83d044-4a1e-4d35-a1ca-cefdd159ac66 tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Deleting the datastore file [datastore2] db09d1f5-88cc-4dc7-9a7b-5d53d09567fa {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 810.257898] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-53b31129-f47d-442b-97d8-6e5260783e39 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.264080] env[62460]: DEBUG oslo_vmware.api [None req-9c83d044-4a1e-4d35-a1ca-cefdd159ac66 tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Waiting for the task: (returnval){ [ 810.264080] env[62460]: value = "task-1313534" [ 810.264080] env[62460]: _type = "Task" [ 810.264080] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.271781] env[62460]: DEBUG oslo_vmware.api [None req-9c83d044-4a1e-4d35-a1ca-cefdd159ac66 tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Task: {'id': task-1313534, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.504245] env[62460]: DEBUG nova.compute.utils [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 810.507892] env[62460]: DEBUG nova.compute.manager [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 810.507892] env[62460]: DEBUG nova.network.neutron [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 810.526227] env[62460]: DEBUG oslo_concurrency.lockutils [None req-cb805937-3829-46aa-b450-89395080ba46 tempest-ServerShowV247Test-2086523556 tempest-ServerShowV247Test-2086523556-project-member] Lock "4ed90d16-81a6-4dbd-8936-0e137151171f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.727s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.565639] env[62460]: DEBUG nova.policy [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ca8d1bcf6c274f1fa383b77393d60d5a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3f26e25609a64eca836f4d9dcdcfd666', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 810.777253] env[62460]: DEBUG oslo_vmware.api [None req-9c83d044-4a1e-4d35-a1ca-cefdd159ac66 tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Task: {'id': task-1313534, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.492195} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.780891] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c83d044-4a1e-4d35-a1ca-cefdd159ac66 tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 810.780891] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-9c83d044-4a1e-4d35-a1ca-cefdd159ac66 tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Deleted contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 810.780891] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-9c83d044-4a1e-4d35-a1ca-cefdd159ac66 tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 810.780891] env[62460]: INFO nova.compute.manager [None req-9c83d044-4a1e-4d35-a1ca-cefdd159ac66 tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Took 1.11 seconds to destroy the instance on the hypervisor. [ 810.780891] env[62460]: DEBUG oslo.service.loopingcall [None req-9c83d044-4a1e-4d35-a1ca-cefdd159ac66 tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 810.781397] env[62460]: DEBUG nova.compute.manager [-] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 810.781397] env[62460]: DEBUG nova.network.neutron [-] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 810.821206] env[62460]: INFO nova.scheduler.client.report [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] Deleted allocations for instance 6c58f9fe-fb6d-4012-8b69-39e4134996b2 [ 810.899146] env[62460]: DEBUG nova.network.neutron [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Successfully created port: 215f9da2-99ec-4af7-8488-b7185454b285 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 811.008280] env[62460]: DEBUG nova.compute.manager [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 811.071027] env[62460]: DEBUG nova.compute.manager [req-f165c8b8-e7d2-417d-b7d9-33f37b79df99 req-40aa20bc-4c4e-43ca-b3ec-5e8f2ca4cf51 service nova] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Received event network-vif-deleted-213b9859-4858-4268-a921-2a6a5d9e1f8a {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 811.071027] env[62460]: INFO nova.compute.manager [req-f165c8b8-e7d2-417d-b7d9-33f37b79df99 req-40aa20bc-4c4e-43ca-b3ec-5e8f2ca4cf51 service nova] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Neutron deleted interface 213b9859-4858-4268-a921-2a6a5d9e1f8a; detaching it from the instance and deleting it from the info cache [ 811.071027] env[62460]: DEBUG nova.network.neutron [req-f165c8b8-e7d2-417d-b7d9-33f37b79df99 req-40aa20bc-4c4e-43ca-b3ec-5e8f2ca4cf51 service nova] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 811.312895] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06fbdf09-e795-4b5b-bf02-bfff27201fd5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.322319] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-821c995d-2999-4999-88a7-ed5a02d6c59b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.356826] env[62460]: DEBUG oslo_concurrency.lockutils [None req-418bb759-54b2-424b-97c6-e6a1ae6b4ea9 tempest-ServerActionsTestOtherA-1594997806 tempest-ServerActionsTestOtherA-1594997806-project-member] Lock "6c58f9fe-fb6d-4012-8b69-39e4134996b2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 131.762s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.362684] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b78d47f7-30bc-42e1-bc7e-29f3f77fde3d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.371578] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04d4a32a-9f48-4696-a756-d00f5fbbc6b4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.389069] env[62460]: DEBUG nova.compute.provider_tree [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 811.537125] env[62460]: DEBUG nova.network.neutron [-] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 811.574220] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-472aabc8-a34a-4f46-824a-87183b905f9c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.584924] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddfaaa38-baa8-448e-822b-f244bd1d8129 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.607673] env[62460]: DEBUG nova.compute.manager [req-f165c8b8-e7d2-417d-b7d9-33f37b79df99 req-40aa20bc-4c4e-43ca-b3ec-5e8f2ca4cf51 service nova] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Detach interface failed, port_id=213b9859-4858-4268-a921-2a6a5d9e1f8a, reason: Instance db09d1f5-88cc-4dc7-9a7b-5d53d09567fa could not be found. {{(pid=62460) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 811.867179] env[62460]: DEBUG nova.compute.manager [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 811.893442] env[62460]: DEBUG nova.scheduler.client.report [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 812.021524] env[62460]: DEBUG nova.compute.manager [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 812.040500] env[62460]: INFO nova.compute.manager [-] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Took 1.26 seconds to deallocate network for instance. [ 812.047741] env[62460]: DEBUG nova.virt.hardware [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 812.047870] env[62460]: DEBUG nova.virt.hardware [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 812.048935] env[62460]: DEBUG nova.virt.hardware [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 812.048935] env[62460]: DEBUG nova.virt.hardware [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 812.048935] env[62460]: DEBUG nova.virt.hardware [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 812.048935] env[62460]: DEBUG nova.virt.hardware [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 812.048935] env[62460]: DEBUG nova.virt.hardware [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 812.049168] env[62460]: DEBUG nova.virt.hardware [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 812.049168] env[62460]: DEBUG nova.virt.hardware [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 812.049168] env[62460]: DEBUG nova.virt.hardware [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 812.049331] env[62460]: DEBUG nova.virt.hardware [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 812.050181] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-414db747-1219-48a3-a6b1-67bdd1510f49 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.060321] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ead933a-5be8-4525-8fa0-652249769d81 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.400041] env[62460]: DEBUG oslo_concurrency.lockutils [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.400s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.400041] env[62460]: DEBUG nova.compute.manager [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 812.403891] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.403891] env[62460]: DEBUG oslo_concurrency.lockutils [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.078s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 812.554595] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9c83d044-4a1e-4d35-a1ca-cefdd159ac66 tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.629727] env[62460]: DEBUG nova.compute.manager [req-a5446920-7de0-41b1-910b-e816570c5c9f req-58934c2b-e1e9-4346-a355-13ecee3b9722 service nova] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Received event network-vif-plugged-215f9da2-99ec-4af7-8488-b7185454b285 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 812.629976] env[62460]: DEBUG oslo_concurrency.lockutils [req-a5446920-7de0-41b1-910b-e816570c5c9f req-58934c2b-e1e9-4346-a355-13ecee3b9722 service nova] Acquiring lock "72e91f8e-0619-464c-b9bc-d6a14be42cb8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.630161] env[62460]: DEBUG oslo_concurrency.lockutils [req-a5446920-7de0-41b1-910b-e816570c5c9f req-58934c2b-e1e9-4346-a355-13ecee3b9722 service nova] Lock "72e91f8e-0619-464c-b9bc-d6a14be42cb8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 812.630331] env[62460]: DEBUG oslo_concurrency.lockutils [req-a5446920-7de0-41b1-910b-e816570c5c9f req-58934c2b-e1e9-4346-a355-13ecee3b9722 service nova] Lock "72e91f8e-0619-464c-b9bc-d6a14be42cb8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.630496] env[62460]: DEBUG nova.compute.manager [req-a5446920-7de0-41b1-910b-e816570c5c9f req-58934c2b-e1e9-4346-a355-13ecee3b9722 service nova] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] No waiting events found dispatching network-vif-plugged-215f9da2-99ec-4af7-8488-b7185454b285 {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 812.630659] env[62460]: WARNING nova.compute.manager [req-a5446920-7de0-41b1-910b-e816570c5c9f req-58934c2b-e1e9-4346-a355-13ecee3b9722 service nova] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Received unexpected event network-vif-plugged-215f9da2-99ec-4af7-8488-b7185454b285 for instance with vm_state building and task_state spawning. [ 812.864890] env[62460]: DEBUG nova.network.neutron [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Successfully updated port: 215f9da2-99ec-4af7-8488-b7185454b285 {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 812.906020] env[62460]: DEBUG nova.compute.utils [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 812.906020] env[62460]: DEBUG nova.compute.manager [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 812.906940] env[62460]: DEBUG nova.network.neutron [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 813.012188] env[62460]: DEBUG nova.policy [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ca8d1bcf6c274f1fa383b77393d60d5a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3f26e25609a64eca836f4d9dcdcfd666', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 813.294463] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccd1c061-9f6c-4555-a956-60a42c1f746b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.304716] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3a6ca49-85fe-4a6a-896d-dc636d4683e9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.344548] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c44dd6d2-6d82-4671-88e9-8e8a331c1df6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.352391] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f60dc6f-fbc6-49dd-a1f9-903702cc1b0d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.367391] env[62460]: DEBUG nova.compute.provider_tree [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 813.369106] env[62460]: DEBUG oslo_concurrency.lockutils [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Acquiring lock "refresh_cache-72e91f8e-0619-464c-b9bc-d6a14be42cb8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 813.369164] env[62460]: DEBUG oslo_concurrency.lockutils [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Acquired lock "refresh_cache-72e91f8e-0619-464c-b9bc-d6a14be42cb8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 813.369294] env[62460]: DEBUG nova.network.neutron [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 813.410711] env[62460]: DEBUG nova.compute.manager [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 813.471775] env[62460]: DEBUG nova.network.neutron [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] Successfully created port: 7b34ac0d-3fb1-4b67-8ef9-9a3a64d990db {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 813.875123] env[62460]: DEBUG nova.scheduler.client.report [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 813.945432] env[62460]: DEBUG nova.network.neutron [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 814.222698] env[62460]: DEBUG nova.network.neutron [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Updating instance_info_cache with network_info: [{"id": "215f9da2-99ec-4af7-8488-b7185454b285", "address": "fa:16:3e:de:e8:6f", "network": {"id": "7428294f-e3f7-45f4-b85a-e51d132281d0", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-867124390-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f26e25609a64eca836f4d9dcdcfd666", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap215f9da2-99", "ovs_interfaceid": "215f9da2-99ec-4af7-8488-b7185454b285", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 814.386513] env[62460]: DEBUG oslo_concurrency.lockutils [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.983s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.387237] env[62460]: ERROR nova.compute.manager [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d8f1c890-d4c1-42ce-9093-2d21538c4bec, please check neutron logs for more information. [ 814.387237] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] Traceback (most recent call last): [ 814.387237] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 814.387237] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] self.driver.spawn(context, instance, image_meta, [ 814.387237] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 814.387237] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 814.387237] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 814.387237] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] vm_ref = self.build_virtual_machine(instance, [ 814.387237] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 814.387237] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] vif_infos = vmwarevif.get_vif_info(self._session, [ 814.387237] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 814.387570] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] for vif in network_info: [ 814.387570] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 814.387570] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] return self._sync_wrapper(fn, *args, **kwargs) [ 814.387570] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 814.387570] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] self.wait() [ 814.387570] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 814.387570] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] self[:] = self._gt.wait() [ 814.387570] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 814.387570] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] return self._exit_event.wait() [ 814.387570] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 814.387570] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] current.throw(*self._exc) [ 814.387570] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 814.387570] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] result = function(*args, **kwargs) [ 814.387870] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 814.387870] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] return func(*args, **kwargs) [ 814.387870] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 814.387870] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] raise e [ 814.387870] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 814.387870] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] nwinfo = self.network_api.allocate_for_instance( [ 814.387870] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 814.387870] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] created_port_ids = self._update_ports_for_instance( [ 814.387870] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 814.387870] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] with excutils.save_and_reraise_exception(): [ 814.387870] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 814.387870] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] self.force_reraise() [ 814.387870] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 814.388356] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] raise self.value [ 814.388356] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 814.388356] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] updated_port = self._update_port( [ 814.388356] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 814.388356] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] _ensure_no_port_binding_failure(port) [ 814.388356] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 814.388356] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] raise exception.PortBindingFailed(port_id=port['id']) [ 814.388356] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] nova.exception.PortBindingFailed: Binding failed for port d8f1c890-d4c1-42ce-9093-2d21538c4bec, please check neutron logs for more information. [ 814.388356] env[62460]: ERROR nova.compute.manager [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] [ 814.388356] env[62460]: DEBUG nova.compute.utils [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] Binding failed for port d8f1c890-d4c1-42ce-9093-2d21538c4bec, please check neutron logs for more information. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 814.390117] env[62460]: DEBUG oslo_concurrency.lockutils [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.068s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.391310] env[62460]: INFO nova.compute.claims [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 814.394373] env[62460]: DEBUG nova.compute.manager [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] Build of instance c7d51b39-b449-4af3-a4d0-c746983ded3e was re-scheduled: Binding failed for port d8f1c890-d4c1-42ce-9093-2d21538c4bec, please check neutron logs for more information. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 814.395029] env[62460]: DEBUG nova.compute.manager [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 814.395116] env[62460]: DEBUG oslo_concurrency.lockutils [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Acquiring lock "refresh_cache-c7d51b39-b449-4af3-a4d0-c746983ded3e" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.395247] env[62460]: DEBUG oslo_concurrency.lockutils [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Acquired lock "refresh_cache-c7d51b39-b449-4af3-a4d0-c746983ded3e" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.395413] env[62460]: DEBUG nova.network.neutron [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 814.426479] env[62460]: DEBUG nova.compute.manager [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 814.463170] env[62460]: DEBUG nova.virt.hardware [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 814.463170] env[62460]: DEBUG nova.virt.hardware [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 814.463276] env[62460]: DEBUG nova.virt.hardware [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 814.463442] env[62460]: DEBUG nova.virt.hardware [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 814.463604] env[62460]: DEBUG nova.virt.hardware [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 814.463752] env[62460]: DEBUG nova.virt.hardware [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 814.463987] env[62460]: DEBUG nova.virt.hardware [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 814.464137] env[62460]: DEBUG nova.virt.hardware [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 814.464302] env[62460]: DEBUG nova.virt.hardware [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 814.464462] env[62460]: DEBUG nova.virt.hardware [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 814.464706] env[62460]: DEBUG nova.virt.hardware [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 814.465593] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92fcf054-6645-4396-9d5c-dd28d64bc868 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.474791] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-017eb00d-39ad-4c2c-b1d6-0474eed6064c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.683486] env[62460]: DEBUG nova.compute.manager [req-de77162e-af3a-4850-8ed9-7ea0a4fa7911 req-7c9be9ad-a05c-47d8-8df6-f9f1312f1894 service nova] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Received event network-changed-215f9da2-99ec-4af7-8488-b7185454b285 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 814.683638] env[62460]: DEBUG nova.compute.manager [req-de77162e-af3a-4850-8ed9-7ea0a4fa7911 req-7c9be9ad-a05c-47d8-8df6-f9f1312f1894 service nova] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Refreshing instance network info cache due to event network-changed-215f9da2-99ec-4af7-8488-b7185454b285. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 814.683837] env[62460]: DEBUG oslo_concurrency.lockutils [req-de77162e-af3a-4850-8ed9-7ea0a4fa7911 req-7c9be9ad-a05c-47d8-8df6-f9f1312f1894 service nova] Acquiring lock "refresh_cache-72e91f8e-0619-464c-b9bc-d6a14be42cb8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.728735] env[62460]: DEBUG oslo_concurrency.lockutils [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Releasing lock "refresh_cache-72e91f8e-0619-464c-b9bc-d6a14be42cb8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 814.728735] env[62460]: DEBUG nova.compute.manager [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Instance network_info: |[{"id": "215f9da2-99ec-4af7-8488-b7185454b285", "address": "fa:16:3e:de:e8:6f", "network": {"id": "7428294f-e3f7-45f4-b85a-e51d132281d0", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-867124390-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f26e25609a64eca836f4d9dcdcfd666", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap215f9da2-99", "ovs_interfaceid": "215f9da2-99ec-4af7-8488-b7185454b285", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 814.729026] env[62460]: DEBUG oslo_concurrency.lockutils [req-de77162e-af3a-4850-8ed9-7ea0a4fa7911 req-7c9be9ad-a05c-47d8-8df6-f9f1312f1894 service nova] Acquired lock "refresh_cache-72e91f8e-0619-464c-b9bc-d6a14be42cb8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.729190] env[62460]: DEBUG nova.network.neutron [req-de77162e-af3a-4850-8ed9-7ea0a4fa7911 req-7c9be9ad-a05c-47d8-8df6-f9f1312f1894 service nova] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Refreshing network info cache for port 215f9da2-99ec-4af7-8488-b7185454b285 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 814.730486] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:de:e8:6f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '03ac2c9c-6ad2-4a85-bfab-c7e336df859a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '215f9da2-99ec-4af7-8488-b7185454b285', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 814.738282] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Creating folder: Project (3f26e25609a64eca836f4d9dcdcfd666). Parent ref: group-v281134. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 814.739165] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7078cee9-5c90-4e6c-b679-7f541445c2b7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.752035] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Created folder: Project (3f26e25609a64eca836f4d9dcdcfd666) in parent group-v281134. [ 814.752035] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Creating folder: Instances. Parent ref: group-v281172. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 814.752035] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-744d331f-8867-40ba-a2c3-d0543c8bcbf9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.762545] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Created folder: Instances in parent group-v281172. [ 814.762824] env[62460]: DEBUG oslo.service.loopingcall [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 814.763031] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 814.763250] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4f362d6d-49c8-49ab-be48-d0c45b4c0a16 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.784474] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 814.784474] env[62460]: value = "task-1313537" [ 814.784474] env[62460]: _type = "Task" [ 814.784474] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.792828] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313537, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.930832] env[62460]: DEBUG nova.network.neutron [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 815.023932] env[62460]: DEBUG nova.network.neutron [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 815.297040] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313537, 'name': CreateVM_Task, 'duration_secs': 0.299749} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.297258] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 815.297893] env[62460]: DEBUG oslo_concurrency.lockutils [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.298088] env[62460]: DEBUG oslo_concurrency.lockutils [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.298513] env[62460]: DEBUG oslo_concurrency.lockutils [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 815.298630] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c6a89534-6982-4edb-b4a5-ca00961e2cbd {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.303299] env[62460]: DEBUG oslo_vmware.api [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Waiting for the task: (returnval){ [ 815.303299] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52df8888-accb-b352-f076-7042e96a31e3" [ 815.303299] env[62460]: _type = "Task" [ 815.303299] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.312602] env[62460]: DEBUG oslo_vmware.api [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52df8888-accb-b352-f076-7042e96a31e3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.435412] env[62460]: DEBUG nova.network.neutron [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] Successfully updated port: 7b34ac0d-3fb1-4b67-8ef9-9a3a64d990db {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 815.526657] env[62460]: DEBUG oslo_concurrency.lockutils [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Releasing lock "refresh_cache-c7d51b39-b449-4af3-a4d0-c746983ded3e" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.526892] env[62460]: DEBUG nova.compute.manager [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 815.527663] env[62460]: DEBUG nova.compute.manager [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 815.527943] env[62460]: DEBUG nova.network.neutron [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 815.554247] env[62460]: DEBUG nova.network.neutron [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 815.592514] env[62460]: DEBUG nova.network.neutron [req-de77162e-af3a-4850-8ed9-7ea0a4fa7911 req-7c9be9ad-a05c-47d8-8df6-f9f1312f1894 service nova] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Updated VIF entry in instance network info cache for port 215f9da2-99ec-4af7-8488-b7185454b285. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 815.592854] env[62460]: DEBUG nova.network.neutron [req-de77162e-af3a-4850-8ed9-7ea0a4fa7911 req-7c9be9ad-a05c-47d8-8df6-f9f1312f1894 service nova] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Updating instance_info_cache with network_info: [{"id": "215f9da2-99ec-4af7-8488-b7185454b285", "address": "fa:16:3e:de:e8:6f", "network": {"id": "7428294f-e3f7-45f4-b85a-e51d132281d0", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-867124390-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f26e25609a64eca836f4d9dcdcfd666", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap215f9da2-99", "ovs_interfaceid": "215f9da2-99ec-4af7-8488-b7185454b285", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 815.815187] env[62460]: DEBUG oslo_vmware.api [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52df8888-accb-b352-f076-7042e96a31e3, 'name': SearchDatastore_Task, 'duration_secs': 0.011894} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.816491] env[62460]: DEBUG oslo_concurrency.lockutils [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.816491] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 815.816491] env[62460]: DEBUG oslo_concurrency.lockutils [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.816491] env[62460]: DEBUG oslo_concurrency.lockutils [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.816720] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 815.817344] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bb6af53d-d46f-4a92-b2bf-805566041cc3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.819683] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efbf6ee8-ecd0-41be-8611-54ed8f87e3d2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.827772] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e22f906-456e-4f58-83fb-2a76c6ca43cf {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.831103] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 815.831299] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 815.832355] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-72759aaa-7864-4df2-8748-4437120cb958 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.863489] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52b6a578-dcb0-4b0a-b256-a9920641db10 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.864583] env[62460]: DEBUG oslo_vmware.api [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Waiting for the task: (returnval){ [ 815.864583] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52c1c2f7-7245-9b73-9ab9-51e93a5930d2" [ 815.864583] env[62460]: _type = "Task" [ 815.864583] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.872690] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4def3464-0723-41be-bb10-620fe3676613 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.880098] env[62460]: DEBUG oslo_vmware.api [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52c1c2f7-7245-9b73-9ab9-51e93a5930d2, 'name': SearchDatastore_Task, 'duration_secs': 0.009016} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.881354] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-97dcb880-c7e2-4b90-b4ba-c405a248dff7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.891835] env[62460]: DEBUG nova.compute.provider_tree [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 815.896501] env[62460]: DEBUG oslo_vmware.api [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Waiting for the task: (returnval){ [ 815.896501] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]5269ac94-7688-15c9-8e3a-42fbfa6ab49b" [ 815.896501] env[62460]: _type = "Task" [ 815.896501] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.905083] env[62460]: DEBUG oslo_vmware.api [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5269ac94-7688-15c9-8e3a-42fbfa6ab49b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.939135] env[62460]: DEBUG oslo_concurrency.lockutils [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Acquiring lock "refresh_cache-3ada3516-3147-4566-a46a-1cb29cf880d0" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.939292] env[62460]: DEBUG oslo_concurrency.lockutils [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Acquired lock "refresh_cache-3ada3516-3147-4566-a46a-1cb29cf880d0" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.939447] env[62460]: DEBUG nova.network.neutron [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 816.056976] env[62460]: DEBUG nova.network.neutron [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.099399] env[62460]: DEBUG oslo_concurrency.lockutils [req-de77162e-af3a-4850-8ed9-7ea0a4fa7911 req-7c9be9ad-a05c-47d8-8df6-f9f1312f1894 service nova] Releasing lock "refresh_cache-72e91f8e-0619-464c-b9bc-d6a14be42cb8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.400448] env[62460]: DEBUG nova.scheduler.client.report [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 816.417934] env[62460]: DEBUG oslo_vmware.api [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5269ac94-7688-15c9-8e3a-42fbfa6ab49b, 'name': SearchDatastore_Task, 'duration_secs': 0.00914} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.419469] env[62460]: DEBUG oslo_concurrency.lockutils [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.419469] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] 72e91f8e-0619-464c-b9bc-d6a14be42cb8/72e91f8e-0619-464c-b9bc-d6a14be42cb8.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 816.419469] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c6482e02-e952-4697-af8e-ea5027bfbc93 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.426995] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 816.427301] env[62460]: DEBUG oslo_vmware.api [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Waiting for the task: (returnval){ [ 816.427301] env[62460]: value = "task-1313538" [ 816.427301] env[62460]: _type = "Task" [ 816.427301] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.427475] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 816.427616] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Starting heal instance info cache {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 816.427733] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Rebuilding the list of instances to heal {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 816.441277] env[62460]: DEBUG oslo_vmware.api [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313538, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.473242] env[62460]: DEBUG nova.network.neutron [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 816.559521] env[62460]: INFO nova.compute.manager [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: c7d51b39-b449-4af3-a4d0-c746983ded3e] Took 1.03 seconds to deallocate network for instance. [ 816.635135] env[62460]: DEBUG nova.network.neutron [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] Updating instance_info_cache with network_info: [{"id": "7b34ac0d-3fb1-4b67-8ef9-9a3a64d990db", "address": "fa:16:3e:d2:02:1e", "network": {"id": "7428294f-e3f7-45f4-b85a-e51d132281d0", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-867124390-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f26e25609a64eca836f4d9dcdcfd666", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7b34ac0d-3f", "ovs_interfaceid": "7b34ac0d-3fb1-4b67-8ef9-9a3a64d990db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.714589] env[62460]: DEBUG nova.compute.manager [req-788f9d01-1d89-47e6-9a95-a54f11ae19c8 req-f90dd48c-de8c-4284-90bd-aa6bab46cf37 service nova] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] Received event network-vif-plugged-7b34ac0d-3fb1-4b67-8ef9-9a3a64d990db {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 816.714822] env[62460]: DEBUG oslo_concurrency.lockutils [req-788f9d01-1d89-47e6-9a95-a54f11ae19c8 req-f90dd48c-de8c-4284-90bd-aa6bab46cf37 service nova] Acquiring lock "3ada3516-3147-4566-a46a-1cb29cf880d0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.715156] env[62460]: DEBUG oslo_concurrency.lockutils [req-788f9d01-1d89-47e6-9a95-a54f11ae19c8 req-f90dd48c-de8c-4284-90bd-aa6bab46cf37 service nova] Lock "3ada3516-3147-4566-a46a-1cb29cf880d0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.715362] env[62460]: DEBUG oslo_concurrency.lockutils [req-788f9d01-1d89-47e6-9a95-a54f11ae19c8 req-f90dd48c-de8c-4284-90bd-aa6bab46cf37 service nova] Lock "3ada3516-3147-4566-a46a-1cb29cf880d0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.715548] env[62460]: DEBUG nova.compute.manager [req-788f9d01-1d89-47e6-9a95-a54f11ae19c8 req-f90dd48c-de8c-4284-90bd-aa6bab46cf37 service nova] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] No waiting events found dispatching network-vif-plugged-7b34ac0d-3fb1-4b67-8ef9-9a3a64d990db {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 816.715719] env[62460]: WARNING nova.compute.manager [req-788f9d01-1d89-47e6-9a95-a54f11ae19c8 req-f90dd48c-de8c-4284-90bd-aa6bab46cf37 service nova] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] Received unexpected event network-vif-plugged-7b34ac0d-3fb1-4b67-8ef9-9a3a64d990db for instance with vm_state building and task_state spawning. [ 816.715950] env[62460]: DEBUG nova.compute.manager [req-788f9d01-1d89-47e6-9a95-a54f11ae19c8 req-f90dd48c-de8c-4284-90bd-aa6bab46cf37 service nova] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] Received event network-changed-7b34ac0d-3fb1-4b67-8ef9-9a3a64d990db {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 816.716145] env[62460]: DEBUG nova.compute.manager [req-788f9d01-1d89-47e6-9a95-a54f11ae19c8 req-f90dd48c-de8c-4284-90bd-aa6bab46cf37 service nova] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] Refreshing instance network info cache due to event network-changed-7b34ac0d-3fb1-4b67-8ef9-9a3a64d990db. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 816.716321] env[62460]: DEBUG oslo_concurrency.lockutils [req-788f9d01-1d89-47e6-9a95-a54f11ae19c8 req-f90dd48c-de8c-4284-90bd-aa6bab46cf37 service nova] Acquiring lock "refresh_cache-3ada3516-3147-4566-a46a-1cb29cf880d0" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.912312] env[62460]: DEBUG oslo_concurrency.lockutils [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.522s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.912419] env[62460]: DEBUG nova.compute.manager [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 816.915165] env[62460]: DEBUG oslo_concurrency.lockutils [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.235s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.916784] env[62460]: INFO nova.compute.claims [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 816.938599] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Skipping network cache update for instance because it is Building. {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 816.939219] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] Skipping network cache update for instance because it is Building. {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 816.940223] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] Skipping network cache update for instance because it is Building. {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 816.959020] env[62460]: DEBUG oslo_vmware.api [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313538, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.506844} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.959020] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] 72e91f8e-0619-464c-b9bc-d6a14be42cb8/72e91f8e-0619-464c-b9bc-d6a14be42cb8.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 816.959020] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 816.959020] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-426b283a-9353-4caf-bb1f-3fdd546afc3b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.962946] env[62460]: DEBUG oslo_vmware.api [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Waiting for the task: (returnval){ [ 816.962946] env[62460]: value = "task-1313539" [ 816.962946] env[62460]: _type = "Task" [ 816.962946] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.968154] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Acquiring lock "refresh_cache-51718896-f5bf-43a9-9396-1ac768737ba2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.968312] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Acquired lock "refresh_cache-51718896-f5bf-43a9-9396-1ac768737ba2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.968428] env[62460]: DEBUG nova.network.neutron [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Forcefully refreshing network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 816.968581] env[62460]: DEBUG nova.objects.instance [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lazy-loading 'info_cache' on Instance uuid 51718896-f5bf-43a9-9396-1ac768737ba2 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 816.975555] env[62460]: DEBUG oslo_vmware.api [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313539, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.138323] env[62460]: DEBUG oslo_concurrency.lockutils [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Releasing lock "refresh_cache-3ada3516-3147-4566-a46a-1cb29cf880d0" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.138632] env[62460]: DEBUG nova.compute.manager [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] Instance network_info: |[{"id": "7b34ac0d-3fb1-4b67-8ef9-9a3a64d990db", "address": "fa:16:3e:d2:02:1e", "network": {"id": "7428294f-e3f7-45f4-b85a-e51d132281d0", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-867124390-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f26e25609a64eca836f4d9dcdcfd666", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7b34ac0d-3f", "ovs_interfaceid": "7b34ac0d-3fb1-4b67-8ef9-9a3a64d990db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 817.139269] env[62460]: DEBUG oslo_concurrency.lockutils [req-788f9d01-1d89-47e6-9a95-a54f11ae19c8 req-f90dd48c-de8c-4284-90bd-aa6bab46cf37 service nova] Acquired lock "refresh_cache-3ada3516-3147-4566-a46a-1cb29cf880d0" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.139361] env[62460]: DEBUG nova.network.neutron [req-788f9d01-1d89-47e6-9a95-a54f11ae19c8 req-f90dd48c-de8c-4284-90bd-aa6bab46cf37 service nova] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] Refreshing network info cache for port 7b34ac0d-3fb1-4b67-8ef9-9a3a64d990db {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 817.140689] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d2:02:1e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '03ac2c9c-6ad2-4a85-bfab-c7e336df859a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7b34ac0d-3fb1-4b67-8ef9-9a3a64d990db', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 817.150039] env[62460]: DEBUG oslo.service.loopingcall [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 817.150449] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 817.151575] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2caf80d3-8b9e-4ad9-bb7e-ac8b6620a8d3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.177088] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 817.177088] env[62460]: value = "task-1313540" [ 817.177088] env[62460]: _type = "Task" [ 817.177088] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.186041] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313540, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.422135] env[62460]: DEBUG nova.compute.utils [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 817.426579] env[62460]: DEBUG nova.compute.manager [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 817.426579] env[62460]: DEBUG nova.network.neutron [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 817.477019] env[62460]: DEBUG oslo_vmware.api [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313539, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064245} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.477388] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 817.478599] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ca58367-6a56-4024-a2da-f1d4d317b359 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.502297] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Reconfiguring VM instance instance-0000003d to attach disk [datastore2] 72e91f8e-0619-464c-b9bc-d6a14be42cb8/72e91f8e-0619-464c-b9bc-d6a14be42cb8.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 817.504623] env[62460]: DEBUG nova.policy [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ca8d1bcf6c274f1fa383b77393d60d5a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3f26e25609a64eca836f4d9dcdcfd666', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 817.506242] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4de9913a-f23e-4aeb-b5fb-680b872d2b40 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.527082] env[62460]: DEBUG oslo_vmware.api [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Waiting for the task: (returnval){ [ 817.527082] env[62460]: value = "task-1313541" [ 817.527082] env[62460]: _type = "Task" [ 817.527082] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.536674] env[62460]: DEBUG oslo_vmware.api [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313541, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.613786] env[62460]: INFO nova.scheduler.client.report [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Deleted allocations for instance c7d51b39-b449-4af3-a4d0-c746983ded3e [ 817.648934] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Acquiring lock "0f5457a0-c918-4407-9ba9-bfeeb949c6a7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.649170] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Lock "0f5457a0-c918-4407-9ba9-bfeeb949c6a7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.695490] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313540, 'name': CreateVM_Task} progress is 99%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.837788] env[62460]: DEBUG nova.network.neutron [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] Successfully created port: 91d45497-1367-4747-acd1-984bd67bc127 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 817.891276] env[62460]: DEBUG nova.network.neutron [req-788f9d01-1d89-47e6-9a95-a54f11ae19c8 req-f90dd48c-de8c-4284-90bd-aa6bab46cf37 service nova] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] Updated VIF entry in instance network info cache for port 7b34ac0d-3fb1-4b67-8ef9-9a3a64d990db. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 817.891712] env[62460]: DEBUG nova.network.neutron [req-788f9d01-1d89-47e6-9a95-a54f11ae19c8 req-f90dd48c-de8c-4284-90bd-aa6bab46cf37 service nova] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] Updating instance_info_cache with network_info: [{"id": "7b34ac0d-3fb1-4b67-8ef9-9a3a64d990db", "address": "fa:16:3e:d2:02:1e", "network": {"id": "7428294f-e3f7-45f4-b85a-e51d132281d0", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-867124390-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f26e25609a64eca836f4d9dcdcfd666", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7b34ac0d-3f", "ovs_interfaceid": "7b34ac0d-3fb1-4b67-8ef9-9a3a64d990db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.926880] env[62460]: DEBUG nova.compute.manager [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 817.996341] env[62460]: DEBUG nova.network.neutron [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 818.040303] env[62460]: DEBUG oslo_vmware.api [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313541, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.122970] env[62460]: DEBUG oslo_concurrency.lockutils [None req-491f0ee7-8e4c-4084-8ad2-e6b12a5d8749 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Lock "c7d51b39-b449-4af3-a4d0-c746983ded3e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 135.818s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.196575] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313540, 'name': CreateVM_Task, 'duration_secs': 0.532152} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.196753] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 818.197936] env[62460]: DEBUG oslo_concurrency.lockutils [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.198203] env[62460]: DEBUG oslo_concurrency.lockutils [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.198752] env[62460]: DEBUG oslo_concurrency.lockutils [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 818.199775] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4024d403-d253-4764-af5c-ed2e0f73a715 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.206398] env[62460]: DEBUG oslo_vmware.api [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Waiting for the task: (returnval){ [ 818.206398] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52f6d636-a8f4-993a-0665-5320dc7b755f" [ 818.206398] env[62460]: _type = "Task" [ 818.206398] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.215851] env[62460]: DEBUG oslo_vmware.api [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52f6d636-a8f4-993a-0665-5320dc7b755f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.275158] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39c8cf39-edb3-4d20-a2e3-4ef5e0d47d52 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.287120] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76ef95d8-dda6-497f-aaed-076b46e4615d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.318821] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cddec83b-c653-4b02-ba9d-6f7971126622 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.326369] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0ce22c2-af23-4e64-bcb3-749cdf4b3032 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.344748] env[62460]: DEBUG nova.compute.provider_tree [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 818.394704] env[62460]: DEBUG oslo_concurrency.lockutils [req-788f9d01-1d89-47e6-9a95-a54f11ae19c8 req-f90dd48c-de8c-4284-90bd-aa6bab46cf37 service nova] Releasing lock "refresh_cache-3ada3516-3147-4566-a46a-1cb29cf880d0" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.537576] env[62460]: DEBUG oslo_vmware.api [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313541, 'name': ReconfigVM_Task, 'duration_secs': 0.650312} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.537864] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Reconfigured VM instance instance-0000003d to attach disk [datastore2] 72e91f8e-0619-464c-b9bc-d6a14be42cb8/72e91f8e-0619-464c-b9bc-d6a14be42cb8.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 818.538586] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-09100020-ea17-4a66-8258-3f2219bb513d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.545134] env[62460]: DEBUG oslo_vmware.api [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Waiting for the task: (returnval){ [ 818.545134] env[62460]: value = "task-1313542" [ 818.545134] env[62460]: _type = "Task" [ 818.545134] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.552819] env[62460]: DEBUG oslo_vmware.api [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313542, 'name': Rename_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.603503] env[62460]: DEBUG nova.network.neutron [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.626442] env[62460]: DEBUG nova.compute.manager [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 818.718432] env[62460]: DEBUG oslo_vmware.api [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52f6d636-a8f4-993a-0665-5320dc7b755f, 'name': SearchDatastore_Task, 'duration_secs': 0.010642} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.718847] env[62460]: DEBUG oslo_concurrency.lockutils [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.719168] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 818.719455] env[62460]: DEBUG oslo_concurrency.lockutils [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.719648] env[62460]: DEBUG oslo_concurrency.lockutils [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.719893] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 818.720405] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7a84f6a9-4649-44bf-846c-835c52a894b0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.728978] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 818.729180] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 818.729916] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c8f7d90b-27a6-42a6-bd94-0f2f4591acc9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.735078] env[62460]: DEBUG oslo_vmware.api [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Waiting for the task: (returnval){ [ 818.735078] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]528f0970-a73f-a643-a191-179a6fb2bc41" [ 818.735078] env[62460]: _type = "Task" [ 818.735078] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.742803] env[62460]: DEBUG oslo_vmware.api [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]528f0970-a73f-a643-a191-179a6fb2bc41, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.850018] env[62460]: DEBUG nova.scheduler.client.report [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 818.941055] env[62460]: DEBUG nova.compute.manager [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 818.970905] env[62460]: DEBUG nova.virt.hardware [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 818.972925] env[62460]: DEBUG nova.virt.hardware [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 818.972925] env[62460]: DEBUG nova.virt.hardware [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 818.972925] env[62460]: DEBUG nova.virt.hardware [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 818.972925] env[62460]: DEBUG nova.virt.hardware [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 818.972925] env[62460]: DEBUG nova.virt.hardware [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 818.973216] env[62460]: DEBUG nova.virt.hardware [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 818.973216] env[62460]: DEBUG nova.virt.hardware [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 818.973216] env[62460]: DEBUG nova.virt.hardware [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 818.973216] env[62460]: DEBUG nova.virt.hardware [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 818.973216] env[62460]: DEBUG nova.virt.hardware [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 818.974341] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a22cc05d-fc2d-4f3e-8de8-62fc3f9aaa12 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.983168] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f47c7fb2-7fe6-48b0-b37f-b8e4e5586ccd {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.054706] env[62460]: DEBUG oslo_vmware.api [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313542, 'name': Rename_Task, 'duration_secs': 0.156933} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.054829] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 819.055159] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8ed209fa-bd59-476c-8e40-a1ec9dc4c798 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.061295] env[62460]: DEBUG oslo_vmware.api [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Waiting for the task: (returnval){ [ 819.061295] env[62460]: value = "task-1313543" [ 819.061295] env[62460]: _type = "Task" [ 819.061295] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.069914] env[62460]: DEBUG oslo_vmware.api [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313543, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.106745] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Releasing lock "refresh_cache-51718896-f5bf-43a9-9396-1ac768737ba2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.107055] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Updated the network info_cache for instance {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 819.107723] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 819.107979] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 819.108222] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 819.108445] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 819.108657] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 819.108858] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 819.109010] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62460) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 819.109175] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 819.156539] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.249068] env[62460]: DEBUG oslo_vmware.api [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]528f0970-a73f-a643-a191-179a6fb2bc41, 'name': SearchDatastore_Task, 'duration_secs': 0.008014} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.250139] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-27653f6c-85db-4774-8d15-84ab83ebc4c7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.256795] env[62460]: DEBUG oslo_vmware.api [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Waiting for the task: (returnval){ [ 819.256795] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52c3c832-59a2-9cda-b571-6d735be29976" [ 819.256795] env[62460]: _type = "Task" [ 819.256795] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.265061] env[62460]: DEBUG oslo_vmware.api [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52c3c832-59a2-9cda-b571-6d735be29976, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.355259] env[62460]: DEBUG oslo_concurrency.lockutils [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.440s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.355960] env[62460]: DEBUG nova.compute.manager [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 819.358644] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.247s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.361545] env[62460]: INFO nova.compute.claims [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 819.571546] env[62460]: DEBUG oslo_vmware.api [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313543, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.613018] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.710713] env[62460]: DEBUG nova.compute.manager [req-dddf4eca-1b05-43df-95bb-51a6054fc7d9 req-ee77e641-c0eb-48a4-9560-6e3638cb0e30 service nova] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] Received event network-vif-plugged-91d45497-1367-4747-acd1-984bd67bc127 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 819.710943] env[62460]: DEBUG oslo_concurrency.lockutils [req-dddf4eca-1b05-43df-95bb-51a6054fc7d9 req-ee77e641-c0eb-48a4-9560-6e3638cb0e30 service nova] Acquiring lock "fde12685-d3b9-46a0-8931-25b904d4f21e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.711179] env[62460]: DEBUG oslo_concurrency.lockutils [req-dddf4eca-1b05-43df-95bb-51a6054fc7d9 req-ee77e641-c0eb-48a4-9560-6e3638cb0e30 service nova] Lock "fde12685-d3b9-46a0-8931-25b904d4f21e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.711350] env[62460]: DEBUG oslo_concurrency.lockutils [req-dddf4eca-1b05-43df-95bb-51a6054fc7d9 req-ee77e641-c0eb-48a4-9560-6e3638cb0e30 service nova] Lock "fde12685-d3b9-46a0-8931-25b904d4f21e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.711520] env[62460]: DEBUG nova.compute.manager [req-dddf4eca-1b05-43df-95bb-51a6054fc7d9 req-ee77e641-c0eb-48a4-9560-6e3638cb0e30 service nova] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] No waiting events found dispatching network-vif-plugged-91d45497-1367-4747-acd1-984bd67bc127 {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 819.711689] env[62460]: WARNING nova.compute.manager [req-dddf4eca-1b05-43df-95bb-51a6054fc7d9 req-ee77e641-c0eb-48a4-9560-6e3638cb0e30 service nova] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] Received unexpected event network-vif-plugged-91d45497-1367-4747-acd1-984bd67bc127 for instance with vm_state building and task_state spawning. [ 819.767961] env[62460]: DEBUG oslo_vmware.api [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52c3c832-59a2-9cda-b571-6d735be29976, 'name': SearchDatastore_Task, 'duration_secs': 0.00871} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.768280] env[62460]: DEBUG oslo_concurrency.lockutils [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.768544] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] 3ada3516-3147-4566-a46a-1cb29cf880d0/3ada3516-3147-4566-a46a-1cb29cf880d0.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 819.768822] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8af82b80-9124-49aa-840c-ec419a5bc829 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.775117] env[62460]: DEBUG oslo_vmware.api [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Waiting for the task: (returnval){ [ 819.775117] env[62460]: value = "task-1313544" [ 819.775117] env[62460]: _type = "Task" [ 819.775117] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.783282] env[62460]: DEBUG oslo_vmware.api [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313544, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.850562] env[62460]: DEBUG nova.network.neutron [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] Successfully updated port: 91d45497-1367-4747-acd1-984bd67bc127 {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 819.864525] env[62460]: DEBUG nova.compute.utils [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 819.867771] env[62460]: DEBUG nova.compute.manager [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 819.870108] env[62460]: DEBUG nova.network.neutron [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 819.913553] env[62460]: DEBUG nova.policy [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '92f9cb84e5034be9b85c3d780178ee97', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a03b6b3cf408486abdfed759728586ba', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 820.073645] env[62460]: DEBUG oslo_vmware.api [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313543, 'name': PowerOnVM_Task, 'duration_secs': 0.874391} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.073907] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 820.074866] env[62460]: INFO nova.compute.manager [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Took 8.05 seconds to spawn the instance on the hypervisor. [ 820.074866] env[62460]: DEBUG nova.compute.manager [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 820.075590] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6046ab66-4c30-4370-b347-878ae426246e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.247816] env[62460]: DEBUG nova.network.neutron [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Successfully created port: c2c96f1c-79ab-4faa-a73b-9d11cae9a02f {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 820.287449] env[62460]: DEBUG oslo_vmware.api [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313544, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.482993} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.287723] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] 3ada3516-3147-4566-a46a-1cb29cf880d0/3ada3516-3147-4566-a46a-1cb29cf880d0.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 820.288055] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 820.288224] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-64d5e7db-8a2e-4475-95e1-fde4937e187e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.295481] env[62460]: DEBUG oslo_vmware.api [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Waiting for the task: (returnval){ [ 820.295481] env[62460]: value = "task-1313545" [ 820.295481] env[62460]: _type = "Task" [ 820.295481] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.303650] env[62460]: DEBUG oslo_vmware.api [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313545, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.355349] env[62460]: DEBUG oslo_concurrency.lockutils [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Acquiring lock "refresh_cache-fde12685-d3b9-46a0-8931-25b904d4f21e" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.355349] env[62460]: DEBUG oslo_concurrency.lockutils [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Acquired lock "refresh_cache-fde12685-d3b9-46a0-8931-25b904d4f21e" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.355349] env[62460]: DEBUG nova.network.neutron [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 820.367403] env[62460]: DEBUG nova.compute.manager [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 820.598746] env[62460]: INFO nova.compute.manager [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Took 38.61 seconds to build instance. [ 820.655948] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-427232cd-030f-4543-a88f-7aca5d088a32 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.663939] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-277b9704-b993-463d-8387-8cf31e8882c7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.694709] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d58549c8-531b-48cf-bd16-ccc4a047d74d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.703034] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c19269bd-0068-4dc4-8fd9-301d57be2ce6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.715595] env[62460]: DEBUG nova.compute.provider_tree [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 820.806595] env[62460]: DEBUG oslo_vmware.api [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313545, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.091878} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.806595] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 820.807394] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11a70328-d5a7-4942-82ee-544085fb54a4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.830017] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] Reconfiguring VM instance instance-0000003e to attach disk [datastore2] 3ada3516-3147-4566-a46a-1cb29cf880d0/3ada3516-3147-4566-a46a-1cb29cf880d0.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 820.830017] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-275e895b-8cb1-4969-8416-dc9ab5291c96 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.848964] env[62460]: DEBUG oslo_vmware.api [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Waiting for the task: (returnval){ [ 820.848964] env[62460]: value = "task-1313546" [ 820.848964] env[62460]: _type = "Task" [ 820.848964] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.858654] env[62460]: DEBUG oslo_vmware.api [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313546, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.908359] env[62460]: DEBUG nova.network.neutron [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 821.079083] env[62460]: DEBUG nova.network.neutron [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] Updating instance_info_cache with network_info: [{"id": "91d45497-1367-4747-acd1-984bd67bc127", "address": "fa:16:3e:5c:6c:5a", "network": {"id": "7428294f-e3f7-45f4-b85a-e51d132281d0", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-867124390-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f26e25609a64eca836f4d9dcdcfd666", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91d45497-13", "ovs_interfaceid": "91d45497-1367-4747-acd1-984bd67bc127", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.100926] env[62460]: DEBUG oslo_concurrency.lockutils [None req-94e49959-ff6f-4bfe-aad8-4e7eee8a1991 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Lock "72e91f8e-0619-464c-b9bc-d6a14be42cb8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 107.325s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.218846] env[62460]: DEBUG nova.scheduler.client.report [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 821.359885] env[62460]: DEBUG oslo_vmware.api [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313546, 'name': ReconfigVM_Task, 'duration_secs': 0.290308} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.360268] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] Reconfigured VM instance instance-0000003e to attach disk [datastore2] 3ada3516-3147-4566-a46a-1cb29cf880d0/3ada3516-3147-4566-a46a-1cb29cf880d0.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 821.360911] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1a8c5592-de63-47e5-b535-04178f800753 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.368223] env[62460]: DEBUG oslo_vmware.api [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Waiting for the task: (returnval){ [ 821.368223] env[62460]: value = "task-1313547" [ 821.368223] env[62460]: _type = "Task" [ 821.368223] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.376198] env[62460]: DEBUG oslo_vmware.api [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313547, 'name': Rename_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.379323] env[62460]: DEBUG nova.compute.manager [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 821.406285] env[62460]: DEBUG nova.virt.hardware [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 821.406553] env[62460]: DEBUG nova.virt.hardware [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 821.406718] env[62460]: DEBUG nova.virt.hardware [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 821.406901] env[62460]: DEBUG nova.virt.hardware [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 821.407068] env[62460]: DEBUG nova.virt.hardware [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 821.407227] env[62460]: DEBUG nova.virt.hardware [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 821.407429] env[62460]: DEBUG nova.virt.hardware [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 821.407586] env[62460]: DEBUG nova.virt.hardware [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 821.407754] env[62460]: DEBUG nova.virt.hardware [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 821.407914] env[62460]: DEBUG nova.virt.hardware [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 821.408093] env[62460]: DEBUG nova.virt.hardware [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 821.408954] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d810fa71-5ea7-4198-9f5b-8f211ed0f39e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.417094] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b586abee-5d42-4de2-9c20-c7481e5e17df {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.582015] env[62460]: DEBUG oslo_concurrency.lockutils [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Releasing lock "refresh_cache-fde12685-d3b9-46a0-8931-25b904d4f21e" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.582257] env[62460]: DEBUG nova.compute.manager [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] Instance network_info: |[{"id": "91d45497-1367-4747-acd1-984bd67bc127", "address": "fa:16:3e:5c:6c:5a", "network": {"id": "7428294f-e3f7-45f4-b85a-e51d132281d0", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-867124390-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f26e25609a64eca836f4d9dcdcfd666", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91d45497-13", "ovs_interfaceid": "91d45497-1367-4747-acd1-984bd67bc127", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 821.582683] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5c:6c:5a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '03ac2c9c-6ad2-4a85-bfab-c7e336df859a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '91d45497-1367-4747-acd1-984bd67bc127', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 821.591156] env[62460]: DEBUG oslo.service.loopingcall [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 821.591156] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 821.591156] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-04e7bc94-9512-4fea-81c7-9817ba8d919e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.605084] env[62460]: DEBUG nova.compute.manager [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 821.614396] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 821.614396] env[62460]: value = "task-1313548" [ 821.614396] env[62460]: _type = "Task" [ 821.614396] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.623651] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313548, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.723921] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.365s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.724509] env[62460]: DEBUG nova.compute.manager [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 821.733021] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.908s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.733021] env[62460]: INFO nova.compute.claims [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 821.734093] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Acquiring lock "3e10af70-db52-4d4a-bb92-821a05dcbab6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.734667] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Lock "3e10af70-db52-4d4a-bb92-821a05dcbab6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.752639] env[62460]: DEBUG nova.compute.manager [req-e6232275-43aa-4ac0-b446-5b509bde1c1b req-9525a545-8dc0-496b-93c7-fb702544301d service nova] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] Received event network-changed-91d45497-1367-4747-acd1-984bd67bc127 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 821.752823] env[62460]: DEBUG nova.compute.manager [req-e6232275-43aa-4ac0-b446-5b509bde1c1b req-9525a545-8dc0-496b-93c7-fb702544301d service nova] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] Refreshing instance network info cache due to event network-changed-91d45497-1367-4747-acd1-984bd67bc127. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 821.753047] env[62460]: DEBUG oslo_concurrency.lockutils [req-e6232275-43aa-4ac0-b446-5b509bde1c1b req-9525a545-8dc0-496b-93c7-fb702544301d service nova] Acquiring lock "refresh_cache-fde12685-d3b9-46a0-8931-25b904d4f21e" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.753197] env[62460]: DEBUG oslo_concurrency.lockutils [req-e6232275-43aa-4ac0-b446-5b509bde1c1b req-9525a545-8dc0-496b-93c7-fb702544301d service nova] Acquired lock "refresh_cache-fde12685-d3b9-46a0-8931-25b904d4f21e" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.753356] env[62460]: DEBUG nova.network.neutron [req-e6232275-43aa-4ac0-b446-5b509bde1c1b req-9525a545-8dc0-496b-93c7-fb702544301d service nova] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] Refreshing network info cache for port 91d45497-1367-4747-acd1-984bd67bc127 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 821.881320] env[62460]: DEBUG oslo_vmware.api [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313547, 'name': Rename_Task, 'duration_secs': 0.14675} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.881624] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 821.881880] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c0c6bb3d-8b12-420d-a99b-58e3744b06eb {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.889995] env[62460]: DEBUG oslo_vmware.api [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Waiting for the task: (returnval){ [ 821.889995] env[62460]: value = "task-1313549" [ 821.889995] env[62460]: _type = "Task" [ 821.889995] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.898071] env[62460]: DEBUG oslo_vmware.api [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313549, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.932560] env[62460]: DEBUG nova.network.neutron [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Successfully updated port: c2c96f1c-79ab-4faa-a73b-9d11cae9a02f {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 822.124545] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313548, 'name': CreateVM_Task} progress is 99%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.127785] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.229261] env[62460]: DEBUG nova.compute.utils [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 822.230722] env[62460]: DEBUG nova.compute.manager [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 822.230898] env[62460]: DEBUG nova.network.neutron [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 822.273250] env[62460]: DEBUG nova.policy [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e531b7c3e7544e53b126d1336ac44c99', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bfa5a65b0f614e769de5b3aa77bf869e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 822.400584] env[62460]: DEBUG oslo_vmware.api [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313549, 'name': PowerOnVM_Task, 'duration_secs': 0.485682} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.400927] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 822.401155] env[62460]: INFO nova.compute.manager [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] Took 7.97 seconds to spawn the instance on the hypervisor. [ 822.401244] env[62460]: DEBUG nova.compute.manager [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 822.402053] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6ed6989-d925-4456-a260-9c9844fa0618 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.436961] env[62460]: DEBUG oslo_concurrency.lockutils [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Acquiring lock "refresh_cache-13890351-6091-4b4f-8484-1cdd0c8523b0" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.436961] env[62460]: DEBUG oslo_concurrency.lockutils [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Acquired lock "refresh_cache-13890351-6091-4b4f-8484-1cdd0c8523b0" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.436961] env[62460]: DEBUG nova.network.neutron [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 822.510696] env[62460]: DEBUG nova.network.neutron [req-e6232275-43aa-4ac0-b446-5b509bde1c1b req-9525a545-8dc0-496b-93c7-fb702544301d service nova] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] Updated VIF entry in instance network info cache for port 91d45497-1367-4747-acd1-984bd67bc127. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 822.510696] env[62460]: DEBUG nova.network.neutron [req-e6232275-43aa-4ac0-b446-5b509bde1c1b req-9525a545-8dc0-496b-93c7-fb702544301d service nova] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] Updating instance_info_cache with network_info: [{"id": "91d45497-1367-4747-acd1-984bd67bc127", "address": "fa:16:3e:5c:6c:5a", "network": {"id": "7428294f-e3f7-45f4-b85a-e51d132281d0", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-867124390-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f26e25609a64eca836f4d9dcdcfd666", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91d45497-13", "ovs_interfaceid": "91d45497-1367-4747-acd1-984bd67bc127", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 822.563609] env[62460]: DEBUG nova.network.neutron [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Successfully created port: bcdde0bd-b227-4477-ae73-bebbae418c5e {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 822.625284] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313548, 'name': CreateVM_Task, 'duration_secs': 0.521046} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.625577] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 822.626117] env[62460]: DEBUG oslo_concurrency.lockutils [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.626412] env[62460]: DEBUG oslo_concurrency.lockutils [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.626600] env[62460]: DEBUG oslo_concurrency.lockutils [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 822.626857] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b154e934-7a8a-4975-9ecd-022e19fa8907 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.632138] env[62460]: DEBUG oslo_vmware.api [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Waiting for the task: (returnval){ [ 822.632138] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52566185-8a55-01a7-4156-d959a9ef20d0" [ 822.632138] env[62460]: _type = "Task" [ 822.632138] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.640591] env[62460]: DEBUG oslo_vmware.api [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52566185-8a55-01a7-4156-d959a9ef20d0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.733889] env[62460]: DEBUG nova.compute.manager [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 822.918635] env[62460]: INFO nova.compute.manager [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] Took 37.14 seconds to build instance. [ 823.001249] env[62460]: DEBUG nova.network.neutron [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 823.014235] env[62460]: DEBUG oslo_concurrency.lockutils [req-e6232275-43aa-4ac0-b446-5b509bde1c1b req-9525a545-8dc0-496b-93c7-fb702544301d service nova] Releasing lock "refresh_cache-fde12685-d3b9-46a0-8931-25b904d4f21e" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.027673] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f66b747-f273-4d5d-8fb6-f800e4415148 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.035712] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0baa0982-fb14-4cda-83b8-7856e793852b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.069066] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad3afe8e-efcf-4598-91ac-ed7158b8ebbb {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.080497] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e273b2bd-84a6-4ee2-a98d-44fd42f57bd2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.094748] env[62460]: DEBUG nova.compute.provider_tree [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 823.144521] env[62460]: DEBUG oslo_vmware.api [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52566185-8a55-01a7-4156-d959a9ef20d0, 'name': SearchDatastore_Task, 'duration_secs': 0.011358} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.145460] env[62460]: DEBUG nova.network.neutron [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Updating instance_info_cache with network_info: [{"id": "c2c96f1c-79ab-4faa-a73b-9d11cae9a02f", "address": "fa:16:3e:e6:93:23", "network": {"id": "504d79e0-28d1-4e95-9654-3863e8f67e75", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-933229445-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a03b6b3cf408486abdfed759728586ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ec763be6-4041-4651-8fd7-3820cf0ab86d", "external-id": "nsx-vlan-transportzone-943", "segmentation_id": 943, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2c96f1c-79", "ovs_interfaceid": "c2c96f1c-79ab-4faa-a73b-9d11cae9a02f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.146679] env[62460]: DEBUG oslo_concurrency.lockutils [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.146951] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 823.147829] env[62460]: DEBUG oslo_concurrency.lockutils [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.147829] env[62460]: DEBUG oslo_concurrency.lockutils [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.147829] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 823.148062] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-eafd658e-2828-478c-ab1e-4331c4d593f9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.158606] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 823.158804] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 823.160124] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d7a27a8a-a7e5-4f92-a899-cb1d109d9764 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.167623] env[62460]: DEBUG oslo_vmware.api [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Waiting for the task: (returnval){ [ 823.167623] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52128c4b-fb16-776e-33e4-30ef09780de9" [ 823.167623] env[62460]: _type = "Task" [ 823.167623] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.177469] env[62460]: DEBUG oslo_vmware.api [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52128c4b-fb16-776e-33e4-30ef09780de9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.422370] env[62460]: DEBUG oslo_concurrency.lockutils [None req-484bc176-2820-4a80-96ee-843c32bcad1f tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Lock "3ada3516-3147-4566-a46a-1cb29cf880d0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 109.296s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.598547] env[62460]: DEBUG nova.scheduler.client.report [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 823.649421] env[62460]: DEBUG oslo_concurrency.lockutils [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Releasing lock "refresh_cache-13890351-6091-4b4f-8484-1cdd0c8523b0" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.649749] env[62460]: DEBUG nova.compute.manager [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Instance network_info: |[{"id": "c2c96f1c-79ab-4faa-a73b-9d11cae9a02f", "address": "fa:16:3e:e6:93:23", "network": {"id": "504d79e0-28d1-4e95-9654-3863e8f67e75", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-933229445-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a03b6b3cf408486abdfed759728586ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ec763be6-4041-4651-8fd7-3820cf0ab86d", "external-id": "nsx-vlan-transportzone-943", "segmentation_id": 943, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2c96f1c-79", "ovs_interfaceid": "c2c96f1c-79ab-4faa-a73b-9d11cae9a02f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 823.650172] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e6:93:23', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ec763be6-4041-4651-8fd7-3820cf0ab86d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c2c96f1c-79ab-4faa-a73b-9d11cae9a02f', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 823.658031] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Creating folder: Project (a03b6b3cf408486abdfed759728586ba). Parent ref: group-v281134. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 823.658240] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-41c12eca-3d83-4d83-a950-d28ddb7e1b94 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.672389] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Created folder: Project (a03b6b3cf408486abdfed759728586ba) in parent group-v281134. [ 823.672551] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Creating folder: Instances. Parent ref: group-v281177. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 823.673217] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ab374159-761c-4c06-a5c0-c41d2a111ced {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.679137] env[62460]: DEBUG oslo_vmware.api [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52128c4b-fb16-776e-33e4-30ef09780de9, 'name': SearchDatastore_Task, 'duration_secs': 0.010702} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.679871] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c5a384fa-f392-48ae-bd00-5c733368714a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.683965] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Created folder: Instances in parent group-v281177. [ 823.683965] env[62460]: DEBUG oslo.service.loopingcall [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 823.683965] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 823.684221] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b55d36d9-1b02-4490-b5ef-8278c4a9396a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.701373] env[62460]: DEBUG oslo_vmware.api [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Waiting for the task: (returnval){ [ 823.701373] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52b573d9-31c6-3438-c800-adb498627d6e" [ 823.701373] env[62460]: _type = "Task" [ 823.701373] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.707443] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 823.707443] env[62460]: value = "task-1313552" [ 823.707443] env[62460]: _type = "Task" [ 823.707443] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.710940] env[62460]: DEBUG oslo_vmware.api [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52b573d9-31c6-3438-c800-adb498627d6e, 'name': SearchDatastore_Task, 'duration_secs': 0.00986} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.713875] env[62460]: DEBUG oslo_concurrency.lockutils [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.714209] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] fde12685-d3b9-46a0-8931-25b904d4f21e/fde12685-d3b9-46a0-8931-25b904d4f21e.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 823.714605] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cf16f7d2-f466-478b-8281-9fd7b740de44 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.722375] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313552, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.723825] env[62460]: DEBUG oslo_vmware.api [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Waiting for the task: (returnval){ [ 823.723825] env[62460]: value = "task-1313553" [ 823.723825] env[62460]: _type = "Task" [ 823.723825] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.733389] env[62460]: DEBUG oslo_vmware.api [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313553, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.749028] env[62460]: DEBUG nova.compute.manager [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 823.778397] env[62460]: DEBUG nova.virt.hardware [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 823.778658] env[62460]: DEBUG nova.virt.hardware [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 823.778863] env[62460]: DEBUG nova.virt.hardware [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 823.779052] env[62460]: DEBUG nova.virt.hardware [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 823.779242] env[62460]: DEBUG nova.virt.hardware [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 823.779742] env[62460]: DEBUG nova.virt.hardware [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 823.779742] env[62460]: DEBUG nova.virt.hardware [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 823.779850] env[62460]: DEBUG nova.virt.hardware [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 823.779997] env[62460]: DEBUG nova.virt.hardware [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 823.780213] env[62460]: DEBUG nova.virt.hardware [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 823.780411] env[62460]: DEBUG nova.virt.hardware [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 823.781370] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaa964fa-44db-4bd9-b361-5f78df11b4e4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.790503] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71e4383f-8221-44d8-9b46-53c20fb8150a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.816171] env[62460]: DEBUG nova.compute.manager [req-1cc62c6a-e388-415f-9146-a3bed955b50f req-8c296ef6-b13b-4537-85f0-f494b1f1baf5 service nova] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Received event network-vif-plugged-c2c96f1c-79ab-4faa-a73b-9d11cae9a02f {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 823.816171] env[62460]: DEBUG oslo_concurrency.lockutils [req-1cc62c6a-e388-415f-9146-a3bed955b50f req-8c296ef6-b13b-4537-85f0-f494b1f1baf5 service nova] Acquiring lock "13890351-6091-4b4f-8484-1cdd0c8523b0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.816171] env[62460]: DEBUG oslo_concurrency.lockutils [req-1cc62c6a-e388-415f-9146-a3bed955b50f req-8c296ef6-b13b-4537-85f0-f494b1f1baf5 service nova] Lock "13890351-6091-4b4f-8484-1cdd0c8523b0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.816171] env[62460]: DEBUG oslo_concurrency.lockutils [req-1cc62c6a-e388-415f-9146-a3bed955b50f req-8c296ef6-b13b-4537-85f0-f494b1f1baf5 service nova] Lock "13890351-6091-4b4f-8484-1cdd0c8523b0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.816602] env[62460]: DEBUG nova.compute.manager [req-1cc62c6a-e388-415f-9146-a3bed955b50f req-8c296ef6-b13b-4537-85f0-f494b1f1baf5 service nova] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] No waiting events found dispatching network-vif-plugged-c2c96f1c-79ab-4faa-a73b-9d11cae9a02f {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 823.816826] env[62460]: WARNING nova.compute.manager [req-1cc62c6a-e388-415f-9146-a3bed955b50f req-8c296ef6-b13b-4537-85f0-f494b1f1baf5 service nova] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Received unexpected event network-vif-plugged-c2c96f1c-79ab-4faa-a73b-9d11cae9a02f for instance with vm_state building and task_state spawning. [ 823.817032] env[62460]: DEBUG nova.compute.manager [req-1cc62c6a-e388-415f-9146-a3bed955b50f req-8c296ef6-b13b-4537-85f0-f494b1f1baf5 service nova] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Received event network-changed-c2c96f1c-79ab-4faa-a73b-9d11cae9a02f {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 823.817353] env[62460]: DEBUG nova.compute.manager [req-1cc62c6a-e388-415f-9146-a3bed955b50f req-8c296ef6-b13b-4537-85f0-f494b1f1baf5 service nova] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Refreshing instance network info cache due to event network-changed-c2c96f1c-79ab-4faa-a73b-9d11cae9a02f. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 823.817593] env[62460]: DEBUG oslo_concurrency.lockutils [req-1cc62c6a-e388-415f-9146-a3bed955b50f req-8c296ef6-b13b-4537-85f0-f494b1f1baf5 service nova] Acquiring lock "refresh_cache-13890351-6091-4b4f-8484-1cdd0c8523b0" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.817752] env[62460]: DEBUG oslo_concurrency.lockutils [req-1cc62c6a-e388-415f-9146-a3bed955b50f req-8c296ef6-b13b-4537-85f0-f494b1f1baf5 service nova] Acquired lock "refresh_cache-13890351-6091-4b4f-8484-1cdd0c8523b0" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.817917] env[62460]: DEBUG nova.network.neutron [req-1cc62c6a-e388-415f-9146-a3bed955b50f req-8c296ef6-b13b-4537-85f0-f494b1f1baf5 service nova] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Refreshing network info cache for port c2c96f1c-79ab-4faa-a73b-9d11cae9a02f {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 823.924845] env[62460]: DEBUG nova.compute.manager [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 824.103941] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.374s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.104712] env[62460]: DEBUG nova.compute.manager [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 824.108134] env[62460]: DEBUG oslo_concurrency.lockutils [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 21.483s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 824.108297] env[62460]: DEBUG nova.objects.instance [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62460) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 824.183781] env[62460]: DEBUG nova.network.neutron [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Successfully updated port: bcdde0bd-b227-4477-ae73-bebbae418c5e {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 824.223827] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313552, 'name': CreateVM_Task} progress is 99%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.237068] env[62460]: DEBUG oslo_vmware.api [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313553, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.450198] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.578207] env[62460]: DEBUG nova.network.neutron [req-1cc62c6a-e388-415f-9146-a3bed955b50f req-8c296ef6-b13b-4537-85f0-f494b1f1baf5 service nova] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Updated VIF entry in instance network info cache for port c2c96f1c-79ab-4faa-a73b-9d11cae9a02f. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 824.578574] env[62460]: DEBUG nova.network.neutron [req-1cc62c6a-e388-415f-9146-a3bed955b50f req-8c296ef6-b13b-4537-85f0-f494b1f1baf5 service nova] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Updating instance_info_cache with network_info: [{"id": "c2c96f1c-79ab-4faa-a73b-9d11cae9a02f", "address": "fa:16:3e:e6:93:23", "network": {"id": "504d79e0-28d1-4e95-9654-3863e8f67e75", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-933229445-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a03b6b3cf408486abdfed759728586ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ec763be6-4041-4651-8fd7-3820cf0ab86d", "external-id": "nsx-vlan-transportzone-943", "segmentation_id": 943, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2c96f1c-79", "ovs_interfaceid": "c2c96f1c-79ab-4faa-a73b-9d11cae9a02f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.613272] env[62460]: DEBUG nova.compute.utils [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 824.617668] env[62460]: DEBUG nova.compute.manager [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 824.617668] env[62460]: DEBUG nova.network.neutron [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 824.661144] env[62460]: DEBUG nova.policy [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '45c39828856b472d894767f94f5d83ae', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cfc48e301fb04231b88ec28a062bda4e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 824.688145] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "refresh_cache-b76f37a0-91d0-4a01-9d95-9c6586081175" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.688145] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquired lock "refresh_cache-b76f37a0-91d0-4a01-9d95-9c6586081175" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.688145] env[62460]: DEBUG nova.network.neutron [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 824.721856] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313552, 'name': CreateVM_Task, 'duration_secs': 0.576574} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.722048] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 824.722780] env[62460]: DEBUG oslo_concurrency.lockutils [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.722861] env[62460]: DEBUG oslo_concurrency.lockutils [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.723247] env[62460]: DEBUG oslo_concurrency.lockutils [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 824.723512] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-230c4d14-28e5-4e4a-92a6-6706067524cb {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.736280] env[62460]: DEBUG oslo_vmware.api [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313553, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.736280] env[62460]: DEBUG oslo_vmware.api [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Waiting for the task: (returnval){ [ 824.736280] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52569c8b-0299-09a9-bbb2-9ebe8e157995" [ 824.736280] env[62460]: _type = "Task" [ 824.736280] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.743944] env[62460]: DEBUG oslo_vmware.api [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52569c8b-0299-09a9-bbb2-9ebe8e157995, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.942693] env[62460]: DEBUG nova.network.neutron [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Successfully created port: 3515dec2-bd6c-4f96-aff1-8dd32de6f7d6 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 825.081210] env[62460]: DEBUG oslo_concurrency.lockutils [req-1cc62c6a-e388-415f-9146-a3bed955b50f req-8c296ef6-b13b-4537-85f0-f494b1f1baf5 service nova] Releasing lock "refresh_cache-13890351-6091-4b4f-8484-1cdd0c8523b0" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.118215] env[62460]: DEBUG nova.compute.manager [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 825.121959] env[62460]: DEBUG oslo_concurrency.lockutils [None req-72f63a57-2f4a-4cfc-90f7-d88c8159df58 tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.123147] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.867s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.124549] env[62460]: INFO nova.compute.claims [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 825.236262] env[62460]: DEBUG oslo_vmware.api [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313553, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.241554] env[62460]: DEBUG nova.network.neutron [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 825.252049] env[62460]: DEBUG oslo_vmware.api [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52569c8b-0299-09a9-bbb2-9ebe8e157995, 'name': SearchDatastore_Task, 'duration_secs': 0.059762} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.252365] env[62460]: DEBUG oslo_concurrency.lockutils [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.252806] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 825.253104] env[62460]: DEBUG oslo_concurrency.lockutils [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.253399] env[62460]: DEBUG oslo_concurrency.lockutils [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.253606] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 825.253877] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-33fbf755-88b9-44a7-961d-9e17e338d293 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.264129] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 825.264129] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 825.264561] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c4669141-4ad9-44ae-a716-ba649f317775 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.274540] env[62460]: DEBUG oslo_vmware.api [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Waiting for the task: (returnval){ [ 825.274540] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]5255a992-ca2d-4b5c-0d15-62bf04e096f8" [ 825.274540] env[62460]: _type = "Task" [ 825.274540] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.285038] env[62460]: DEBUG oslo_vmware.api [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5255a992-ca2d-4b5c-0d15-62bf04e096f8, 'name': SearchDatastore_Task, 'duration_secs': 0.011062} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.288052] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fd760e02-1f89-423c-bb9d-19890651dcdf {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.294330] env[62460]: DEBUG oslo_vmware.api [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Waiting for the task: (returnval){ [ 825.294330] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52572772-352e-08d8-ff29-f0b613eb9811" [ 825.294330] env[62460]: _type = "Task" [ 825.294330] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.302611] env[62460]: DEBUG oslo_vmware.api [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52572772-352e-08d8-ff29-f0b613eb9811, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.406494] env[62460]: DEBUG nova.network.neutron [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Updating instance_info_cache with network_info: [{"id": "bcdde0bd-b227-4477-ae73-bebbae418c5e", "address": "fa:16:3e:25:53:33", "network": {"id": "3c49720c-94c9-4e2f-bbb7-4266cc97faee", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1650281790-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfa5a65b0f614e769de5b3aa77bf869e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbcdde0bd-b2", "ovs_interfaceid": "bcdde0bd-b227-4477-ae73-bebbae418c5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 825.737659] env[62460]: DEBUG oslo_vmware.api [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313553, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.546733} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.737945] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] fde12685-d3b9-46a0-8931-25b904d4f21e/fde12685-d3b9-46a0-8931-25b904d4f21e.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 825.738186] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 825.738460] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e4f2fc8c-d51c-4d7e-b18e-d98209d32608 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.746515] env[62460]: DEBUG oslo_vmware.api [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Waiting for the task: (returnval){ [ 825.746515] env[62460]: value = "task-1313554" [ 825.746515] env[62460]: _type = "Task" [ 825.746515] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.755605] env[62460]: DEBUG oslo_vmware.api [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313554, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.807678] env[62460]: DEBUG oslo_vmware.api [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52572772-352e-08d8-ff29-f0b613eb9811, 'name': SearchDatastore_Task, 'duration_secs': 0.009876} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.807971] env[62460]: DEBUG oslo_concurrency.lockutils [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.808592] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] 13890351-6091-4b4f-8484-1cdd0c8523b0/13890351-6091-4b4f-8484-1cdd0c8523b0.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 825.808592] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fe06be28-2e1b-4111-94d0-b678f491e7da {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.818020] env[62460]: DEBUG oslo_vmware.api [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Waiting for the task: (returnval){ [ 825.818020] env[62460]: value = "task-1313555" [ 825.818020] env[62460]: _type = "Task" [ 825.818020] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.825407] env[62460]: DEBUG oslo_vmware.api [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Task: {'id': task-1313555, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.844204] env[62460]: DEBUG nova.compute.manager [req-2f839cd7-c3c2-4297-95d8-6816096cb747 req-fa10ca46-7eee-4264-a05d-6c318b370e76 service nova] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Received event network-vif-plugged-bcdde0bd-b227-4477-ae73-bebbae418c5e {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 825.844450] env[62460]: DEBUG oslo_concurrency.lockutils [req-2f839cd7-c3c2-4297-95d8-6816096cb747 req-fa10ca46-7eee-4264-a05d-6c318b370e76 service nova] Acquiring lock "b76f37a0-91d0-4a01-9d95-9c6586081175-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.844691] env[62460]: DEBUG oslo_concurrency.lockutils [req-2f839cd7-c3c2-4297-95d8-6816096cb747 req-fa10ca46-7eee-4264-a05d-6c318b370e76 service nova] Lock "b76f37a0-91d0-4a01-9d95-9c6586081175-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.844886] env[62460]: DEBUG oslo_concurrency.lockutils [req-2f839cd7-c3c2-4297-95d8-6816096cb747 req-fa10ca46-7eee-4264-a05d-6c318b370e76 service nova] Lock "b76f37a0-91d0-4a01-9d95-9c6586081175-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.845145] env[62460]: DEBUG nova.compute.manager [req-2f839cd7-c3c2-4297-95d8-6816096cb747 req-fa10ca46-7eee-4264-a05d-6c318b370e76 service nova] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] No waiting events found dispatching network-vif-plugged-bcdde0bd-b227-4477-ae73-bebbae418c5e {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 825.845347] env[62460]: WARNING nova.compute.manager [req-2f839cd7-c3c2-4297-95d8-6816096cb747 req-fa10ca46-7eee-4264-a05d-6c318b370e76 service nova] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Received unexpected event network-vif-plugged-bcdde0bd-b227-4477-ae73-bebbae418c5e for instance with vm_state building and task_state spawning. [ 825.845543] env[62460]: DEBUG nova.compute.manager [req-2f839cd7-c3c2-4297-95d8-6816096cb747 req-fa10ca46-7eee-4264-a05d-6c318b370e76 service nova] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Received event network-changed-bcdde0bd-b227-4477-ae73-bebbae418c5e {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 825.845740] env[62460]: DEBUG nova.compute.manager [req-2f839cd7-c3c2-4297-95d8-6816096cb747 req-fa10ca46-7eee-4264-a05d-6c318b370e76 service nova] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Refreshing instance network info cache due to event network-changed-bcdde0bd-b227-4477-ae73-bebbae418c5e. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 825.845946] env[62460]: DEBUG oslo_concurrency.lockutils [req-2f839cd7-c3c2-4297-95d8-6816096cb747 req-fa10ca46-7eee-4264-a05d-6c318b370e76 service nova] Acquiring lock "refresh_cache-b76f37a0-91d0-4a01-9d95-9c6586081175" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.909686] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Releasing lock "refresh_cache-b76f37a0-91d0-4a01-9d95-9c6586081175" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.910132] env[62460]: DEBUG nova.compute.manager [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Instance network_info: |[{"id": "bcdde0bd-b227-4477-ae73-bebbae418c5e", "address": "fa:16:3e:25:53:33", "network": {"id": "3c49720c-94c9-4e2f-bbb7-4266cc97faee", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1650281790-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfa5a65b0f614e769de5b3aa77bf869e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbcdde0bd-b2", "ovs_interfaceid": "bcdde0bd-b227-4477-ae73-bebbae418c5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 825.910605] env[62460]: DEBUG oslo_concurrency.lockutils [req-2f839cd7-c3c2-4297-95d8-6816096cb747 req-fa10ca46-7eee-4264-a05d-6c318b370e76 service nova] Acquired lock "refresh_cache-b76f37a0-91d0-4a01-9d95-9c6586081175" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.910845] env[62460]: DEBUG nova.network.neutron [req-2f839cd7-c3c2-4297-95d8-6816096cb747 req-fa10ca46-7eee-4264-a05d-6c318b370e76 service nova] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Refreshing network info cache for port bcdde0bd-b227-4477-ae73-bebbae418c5e {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 825.912548] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:25:53:33', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '04ccbc7a-cf8d-4ea2-8411-291a1e27df7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bcdde0bd-b227-4477-ae73-bebbae418c5e', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 825.920784] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Creating folder: Project (bfa5a65b0f614e769de5b3aa77bf869e). Parent ref: group-v281134. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 825.921667] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0ac23826-1a66-4d9e-b6df-b1a3081be393 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.933951] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Created folder: Project (bfa5a65b0f614e769de5b3aa77bf869e) in parent group-v281134. [ 825.934220] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Creating folder: Instances. Parent ref: group-v281180. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 825.934476] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bd8044c6-1e9e-49fc-8fa0-2d652b226591 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.945696] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Created folder: Instances in parent group-v281180. [ 825.945943] env[62460]: DEBUG oslo.service.loopingcall [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 825.946169] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 825.946420] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-23b56417-9c1d-4e14-9cc6-38fea39cecbe {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.966904] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 825.966904] env[62460]: value = "task-1313558" [ 825.966904] env[62460]: _type = "Task" [ 825.966904] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.975636] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313558, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.130499] env[62460]: DEBUG nova.compute.manager [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 826.161595] env[62460]: DEBUG nova.virt.hardware [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 826.162092] env[62460]: DEBUG nova.virt.hardware [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 826.162456] env[62460]: DEBUG nova.virt.hardware [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 826.162766] env[62460]: DEBUG nova.virt.hardware [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 826.163036] env[62460]: DEBUG nova.virt.hardware [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 826.163338] env[62460]: DEBUG nova.virt.hardware [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 826.163676] env[62460]: DEBUG nova.virt.hardware [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 826.163964] env[62460]: DEBUG nova.virt.hardware [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 826.164271] env[62460]: DEBUG nova.virt.hardware [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 826.164561] env[62460]: DEBUG nova.virt.hardware [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 826.164918] env[62460]: DEBUG nova.virt.hardware [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 826.166085] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9606f53f-88e7-4ac0-b088-c30d19b73ee6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.180033] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-821ec7a6-4eb1-4b1a-a930-44e764eb1acd {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.263340] env[62460]: DEBUG oslo_vmware.api [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313554, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070627} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.263485] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 826.267459] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b16f5356-2d61-4d8d-84ba-47baa8ffa0f8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.295180] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] fde12685-d3b9-46a0-8931-25b904d4f21e/fde12685-d3b9-46a0-8931-25b904d4f21e.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 826.297399] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-317f706f-3e28-4658-8c67-5a20600f84b5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.322941] env[62460]: DEBUG oslo_vmware.api [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Waiting for the task: (returnval){ [ 826.322941] env[62460]: value = "task-1313559" [ 826.322941] env[62460]: _type = "Task" [ 826.322941] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.332527] env[62460]: DEBUG oslo_vmware.api [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Task: {'id': task-1313555, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.489893} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.333149] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] 13890351-6091-4b4f-8484-1cdd0c8523b0/13890351-6091-4b4f-8484-1cdd0c8523b0.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 826.333491] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 826.333687] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4444c2f0-32f6-4b3a-a132-826db5a04704 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.339362] env[62460]: DEBUG oslo_vmware.api [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313559, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.349357] env[62460]: DEBUG oslo_vmware.api [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Waiting for the task: (returnval){ [ 826.349357] env[62460]: value = "task-1313560" [ 826.349357] env[62460]: _type = "Task" [ 826.349357] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.362792] env[62460]: DEBUG oslo_vmware.api [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Task: {'id': task-1313560, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.477454] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313558, 'name': CreateVM_Task} progress is 99%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.527468] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c017b45d-0ad3-41d7-900d-8591cb13aa4f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.537288] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a543dd3-d5d9-4dcd-986e-5f94905af261 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.569614] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff074089-84ea-4564-bed7-da2266b28456 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.578749] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61024d11-274f-4723-9015-0fb4c17b7779 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.594765] env[62460]: DEBUG nova.compute.provider_tree [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 826.634993] env[62460]: DEBUG nova.network.neutron [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Successfully updated port: 3515dec2-bd6c-4f96-aff1-8dd32de6f7d6 {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 826.709116] env[62460]: DEBUG nova.network.neutron [req-2f839cd7-c3c2-4297-95d8-6816096cb747 req-fa10ca46-7eee-4264-a05d-6c318b370e76 service nova] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Updated VIF entry in instance network info cache for port bcdde0bd-b227-4477-ae73-bebbae418c5e. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 826.709484] env[62460]: DEBUG nova.network.neutron [req-2f839cd7-c3c2-4297-95d8-6816096cb747 req-fa10ca46-7eee-4264-a05d-6c318b370e76 service nova] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Updating instance_info_cache with network_info: [{"id": "bcdde0bd-b227-4477-ae73-bebbae418c5e", "address": "fa:16:3e:25:53:33", "network": {"id": "3c49720c-94c9-4e2f-bbb7-4266cc97faee", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1650281790-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfa5a65b0f614e769de5b3aa77bf869e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbcdde0bd-b2", "ovs_interfaceid": "bcdde0bd-b227-4477-ae73-bebbae418c5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.834706] env[62460]: DEBUG oslo_vmware.api [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313559, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.859738] env[62460]: DEBUG oslo_vmware.api [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Task: {'id': task-1313560, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072299} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.860009] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 826.860767] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88351404-ed34-453a-aa13-b125f0b76110 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.884456] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Reconfiguring VM instance instance-00000040 to attach disk [datastore2] 13890351-6091-4b4f-8484-1cdd0c8523b0/13890351-6091-4b4f-8484-1cdd0c8523b0.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 826.884703] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6668b263-6466-47af-abcf-3834399dfcf0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.904700] env[62460]: DEBUG oslo_vmware.api [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Waiting for the task: (returnval){ [ 826.904700] env[62460]: value = "task-1313561" [ 826.904700] env[62460]: _type = "Task" [ 826.904700] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.912908] env[62460]: DEBUG oslo_vmware.api [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Task: {'id': task-1313561, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.978641] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313558, 'name': CreateVM_Task, 'duration_secs': 0.541795} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.978855] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 826.979539] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.979720] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.980060] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 826.980572] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-27146222-eb81-4868-b030-055cb976bc83 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.985906] env[62460]: DEBUG oslo_vmware.api [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 826.985906] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]5250f806-f799-370c-d7fb-68e7e9dc2c5e" [ 826.985906] env[62460]: _type = "Task" [ 826.985906] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.993698] env[62460]: DEBUG oslo_vmware.api [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5250f806-f799-370c-d7fb-68e7e9dc2c5e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.098727] env[62460]: DEBUG nova.scheduler.client.report [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 827.140022] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Acquiring lock "refresh_cache-5b6e8205-003e-49c0-a73d-be2e032a8272" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.140212] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Acquired lock "refresh_cache-5b6e8205-003e-49c0-a73d-be2e032a8272" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.140423] env[62460]: DEBUG nova.network.neutron [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 827.213403] env[62460]: DEBUG oslo_concurrency.lockutils [req-2f839cd7-c3c2-4297-95d8-6816096cb747 req-fa10ca46-7eee-4264-a05d-6c318b370e76 service nova] Releasing lock "refresh_cache-b76f37a0-91d0-4a01-9d95-9c6586081175" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.335016] env[62460]: DEBUG oslo_vmware.api [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313559, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.414706] env[62460]: DEBUG oslo_vmware.api [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Task: {'id': task-1313561, 'name': ReconfigVM_Task, 'duration_secs': 0.282816} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.415011] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Reconfigured VM instance instance-00000040 to attach disk [datastore2] 13890351-6091-4b4f-8484-1cdd0c8523b0/13890351-6091-4b4f-8484-1cdd0c8523b0.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 827.415675] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6a03c683-1ad3-4f1e-8164-b43f5850477c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.424542] env[62460]: DEBUG oslo_vmware.api [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Waiting for the task: (returnval){ [ 827.424542] env[62460]: value = "task-1313562" [ 827.424542] env[62460]: _type = "Task" [ 827.424542] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.433358] env[62460]: DEBUG oslo_vmware.api [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Task: {'id': task-1313562, 'name': Rename_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.496910] env[62460]: DEBUG oslo_vmware.api [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5250f806-f799-370c-d7fb-68e7e9dc2c5e, 'name': SearchDatastore_Task, 'duration_secs': 0.009845} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.497640] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.497640] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 827.497891] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.497891] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.498080] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 827.498354] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f8cff5eb-d10a-4920-8524-28e78630a45e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.507479] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 827.507656] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 827.508437] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-311989ea-d084-402c-8dae-d81ee613853e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.514478] env[62460]: DEBUG oslo_vmware.api [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 827.514478] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52ad8229-f13b-2f04-c275-96cbb70be795" [ 827.514478] env[62460]: _type = "Task" [ 827.514478] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.522888] env[62460]: DEBUG oslo_vmware.api [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52ad8229-f13b-2f04-c275-96cbb70be795, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.603166] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.480s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.603745] env[62460]: DEBUG nova.compute.manager [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 827.606650] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7b25a346-a87d-486c-86ee-f9dc9c62cc1e tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.800s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.606869] env[62460]: DEBUG nova.objects.instance [None req-7b25a346-a87d-486c-86ee-f9dc9c62cc1e tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Lazy-loading 'resources' on Instance uuid 51718896-f5bf-43a9-9396-1ac768737ba2 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 827.687751] env[62460]: DEBUG nova.network.neutron [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 827.836923] env[62460]: DEBUG oslo_vmware.api [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313559, 'name': ReconfigVM_Task, 'duration_secs': 1.068594} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.837212] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] Reconfigured VM instance instance-0000003f to attach disk [datastore2] fde12685-d3b9-46a0-8931-25b904d4f21e/fde12685-d3b9-46a0-8931-25b904d4f21e.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 827.837619] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8ddc44c1-d0ae-4535-95e4-82960561ddf3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.846586] env[62460]: DEBUG oslo_vmware.api [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Waiting for the task: (returnval){ [ 827.846586] env[62460]: value = "task-1313563" [ 827.846586] env[62460]: _type = "Task" [ 827.846586] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.855614] env[62460]: DEBUG oslo_vmware.api [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313563, 'name': Rename_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.872970] env[62460]: DEBUG nova.compute.manager [req-419bee05-82c2-4230-b431-6607905232a9 req-c5b3fed4-5234-4f89-a7b7-393180fc054c service nova] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Received event network-vif-plugged-3515dec2-bd6c-4f96-aff1-8dd32de6f7d6 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 827.873252] env[62460]: DEBUG oslo_concurrency.lockutils [req-419bee05-82c2-4230-b431-6607905232a9 req-c5b3fed4-5234-4f89-a7b7-393180fc054c service nova] Acquiring lock "5b6e8205-003e-49c0-a73d-be2e032a8272-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.873458] env[62460]: DEBUG oslo_concurrency.lockutils [req-419bee05-82c2-4230-b431-6607905232a9 req-c5b3fed4-5234-4f89-a7b7-393180fc054c service nova] Lock "5b6e8205-003e-49c0-a73d-be2e032a8272-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.873691] env[62460]: DEBUG oslo_concurrency.lockutils [req-419bee05-82c2-4230-b431-6607905232a9 req-c5b3fed4-5234-4f89-a7b7-393180fc054c service nova] Lock "5b6e8205-003e-49c0-a73d-be2e032a8272-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.873798] env[62460]: DEBUG nova.compute.manager [req-419bee05-82c2-4230-b431-6607905232a9 req-c5b3fed4-5234-4f89-a7b7-393180fc054c service nova] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] No waiting events found dispatching network-vif-plugged-3515dec2-bd6c-4f96-aff1-8dd32de6f7d6 {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 827.873964] env[62460]: WARNING nova.compute.manager [req-419bee05-82c2-4230-b431-6607905232a9 req-c5b3fed4-5234-4f89-a7b7-393180fc054c service nova] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Received unexpected event network-vif-plugged-3515dec2-bd6c-4f96-aff1-8dd32de6f7d6 for instance with vm_state building and task_state spawning. [ 827.874777] env[62460]: DEBUG nova.compute.manager [req-419bee05-82c2-4230-b431-6607905232a9 req-c5b3fed4-5234-4f89-a7b7-393180fc054c service nova] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Received event network-changed-3515dec2-bd6c-4f96-aff1-8dd32de6f7d6 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 827.874936] env[62460]: DEBUG nova.compute.manager [req-419bee05-82c2-4230-b431-6607905232a9 req-c5b3fed4-5234-4f89-a7b7-393180fc054c service nova] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Refreshing instance network info cache due to event network-changed-3515dec2-bd6c-4f96-aff1-8dd32de6f7d6. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 827.875223] env[62460]: DEBUG oslo_concurrency.lockutils [req-419bee05-82c2-4230-b431-6607905232a9 req-c5b3fed4-5234-4f89-a7b7-393180fc054c service nova] Acquiring lock "refresh_cache-5b6e8205-003e-49c0-a73d-be2e032a8272" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.892589] env[62460]: DEBUG nova.network.neutron [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Updating instance_info_cache with network_info: [{"id": "3515dec2-bd6c-4f96-aff1-8dd32de6f7d6", "address": "fa:16:3e:53:9c:d5", "network": {"id": "4cb03486-3bea-4c6a-a019-37486f271107", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-2094026403-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfc48e301fb04231b88ec28a062bda4e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c7d2575f-b92f-44ec-a863-634cb76631a2", "external-id": "nsx-vlan-transportzone-794", "segmentation_id": 794, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3515dec2-bd", "ovs_interfaceid": "3515dec2-bd6c-4f96-aff1-8dd32de6f7d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.935851] env[62460]: DEBUG oslo_vmware.api [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Task: {'id': task-1313562, 'name': Rename_Task, 'duration_secs': 0.147681} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.936181] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 827.936445] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4d34e1ca-5db9-400d-a520-b7bdabc4e851 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.957656] env[62460]: DEBUG oslo_vmware.api [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Waiting for the task: (returnval){ [ 827.957656] env[62460]: value = "task-1313564" [ 827.957656] env[62460]: _type = "Task" [ 827.957656] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.026395] env[62460]: DEBUG oslo_vmware.api [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52ad8229-f13b-2f04-c275-96cbb70be795, 'name': SearchDatastore_Task, 'duration_secs': 0.009531} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.027335] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-21e42bb2-2756-47e5-b4ed-bf9d51b2b077 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.033112] env[62460]: DEBUG oslo_vmware.api [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 828.033112] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52e77bcd-feee-881d-bf44-166b1cce5e2e" [ 828.033112] env[62460]: _type = "Task" [ 828.033112] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.041543] env[62460]: DEBUG oslo_vmware.api [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52e77bcd-feee-881d-bf44-166b1cce5e2e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.110328] env[62460]: DEBUG nova.compute.utils [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 828.111795] env[62460]: DEBUG nova.compute.manager [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 828.111941] env[62460]: DEBUG nova.network.neutron [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 828.154292] env[62460]: DEBUG nova.policy [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b4c1b6d1942542a7815aee4ec860259a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '44220d464249478fb1373bc6b146ee0f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 828.359967] env[62460]: DEBUG oslo_vmware.api [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313563, 'name': Rename_Task, 'duration_secs': 0.186242} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.360302] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 828.360556] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cda4b06f-f60d-492c-a5e0-5e9ae8ab256e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.368824] env[62460]: DEBUG oslo_vmware.api [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Waiting for the task: (returnval){ [ 828.368824] env[62460]: value = "task-1313565" [ 828.368824] env[62460]: _type = "Task" [ 828.368824] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.370606] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f080dfb-5210-4863-922e-ca5fb83f2e8c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.383733] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74703145-7ed0-4907-b2ea-3f36e1aa4a9c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.387320] env[62460]: DEBUG oslo_vmware.api [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313565, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.415757] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Releasing lock "refresh_cache-5b6e8205-003e-49c0-a73d-be2e032a8272" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 828.416145] env[62460]: DEBUG nova.compute.manager [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Instance network_info: |[{"id": "3515dec2-bd6c-4f96-aff1-8dd32de6f7d6", "address": "fa:16:3e:53:9c:d5", "network": {"id": "4cb03486-3bea-4c6a-a019-37486f271107", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-2094026403-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfc48e301fb04231b88ec28a062bda4e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c7d2575f-b92f-44ec-a863-634cb76631a2", "external-id": "nsx-vlan-transportzone-794", "segmentation_id": 794, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3515dec2-bd", "ovs_interfaceid": "3515dec2-bd6c-4f96-aff1-8dd32de6f7d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 828.416720] env[62460]: DEBUG oslo_concurrency.lockutils [req-419bee05-82c2-4230-b431-6607905232a9 req-c5b3fed4-5234-4f89-a7b7-393180fc054c service nova] Acquired lock "refresh_cache-5b6e8205-003e-49c0-a73d-be2e032a8272" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.416947] env[62460]: DEBUG nova.network.neutron [req-419bee05-82c2-4230-b431-6607905232a9 req-c5b3fed4-5234-4f89-a7b7-393180fc054c service nova] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Refreshing network info cache for port 3515dec2-bd6c-4f96-aff1-8dd32de6f7d6 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 828.418860] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:53:9c:d5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c7d2575f-b92f-44ec-a863-634cb76631a2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3515dec2-bd6c-4f96-aff1-8dd32de6f7d6', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 828.426850] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Creating folder: Project (cfc48e301fb04231b88ec28a062bda4e). Parent ref: group-v281134. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 828.427613] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5d19efe-b990-42dd-a75a-596e2654b806 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.430963] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0bf673ff-fb59-48ad-8d8e-a744671908d4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.441437] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c8a04d2-f0cd-4d8b-903d-5761a6bf5e79 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.448864] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Created folder: Project (cfc48e301fb04231b88ec28a062bda4e) in parent group-v281134. [ 828.449093] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Creating folder: Instances. Parent ref: group-v281183. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 828.449412] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4b47c65a-28ca-4ef5-80ac-cc8455c0de90 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.466098] env[62460]: DEBUG nova.compute.provider_tree [None req-7b25a346-a87d-486c-86ee-f9dc9c62cc1e tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 828.472755] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Created folder: Instances in parent group-v281183. [ 828.472755] env[62460]: DEBUG oslo.service.loopingcall [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 828.473601] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 828.475435] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dedb898a-134c-4895-a853-bf2290aca0f2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.494208] env[62460]: DEBUG nova.network.neutron [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] Successfully created port: 608a7ef6-983b-4023-a6b6-69e632436554 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 828.496751] env[62460]: DEBUG oslo_vmware.api [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Task: {'id': task-1313564, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.502972] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 828.502972] env[62460]: value = "task-1313568" [ 828.502972] env[62460]: _type = "Task" [ 828.502972] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.511604] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313568, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.543728] env[62460]: DEBUG oslo_vmware.api [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52e77bcd-feee-881d-bf44-166b1cce5e2e, 'name': SearchDatastore_Task, 'duration_secs': 0.033094} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.544016] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 828.544297] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] b76f37a0-91d0-4a01-9d95-9c6586081175/b76f37a0-91d0-4a01-9d95-9c6586081175.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 828.544568] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5a044974-e113-465a-82dc-0f7e9477e771 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.553562] env[62460]: DEBUG oslo_vmware.api [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 828.553562] env[62460]: value = "task-1313569" [ 828.553562] env[62460]: _type = "Task" [ 828.553562] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.562454] env[62460]: DEBUG oslo_vmware.api [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313569, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.617634] env[62460]: DEBUG nova.compute.manager [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 828.882544] env[62460]: DEBUG oslo_vmware.api [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313565, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.973272] env[62460]: DEBUG nova.scheduler.client.report [None req-7b25a346-a87d-486c-86ee-f9dc9c62cc1e tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 828.986709] env[62460]: DEBUG oslo_vmware.api [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Task: {'id': task-1313564, 'name': PowerOnVM_Task, 'duration_secs': 0.546873} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.986709] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 828.986709] env[62460]: INFO nova.compute.manager [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Took 7.61 seconds to spawn the instance on the hypervisor. [ 828.986709] env[62460]: DEBUG nova.compute.manager [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 828.988044] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9907fc46-bcb1-418f-8373-af4eac6056e8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.020938] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313568, 'name': CreateVM_Task} progress is 99%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.066114] env[62460]: DEBUG oslo_vmware.api [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313569, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.219870] env[62460]: DEBUG nova.network.neutron [req-419bee05-82c2-4230-b431-6607905232a9 req-c5b3fed4-5234-4f89-a7b7-393180fc054c service nova] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Updated VIF entry in instance network info cache for port 3515dec2-bd6c-4f96-aff1-8dd32de6f7d6. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 829.220376] env[62460]: DEBUG nova.network.neutron [req-419bee05-82c2-4230-b431-6607905232a9 req-c5b3fed4-5234-4f89-a7b7-393180fc054c service nova] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Updating instance_info_cache with network_info: [{"id": "3515dec2-bd6c-4f96-aff1-8dd32de6f7d6", "address": "fa:16:3e:53:9c:d5", "network": {"id": "4cb03486-3bea-4c6a-a019-37486f271107", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-2094026403-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfc48e301fb04231b88ec28a062bda4e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c7d2575f-b92f-44ec-a863-634cb76631a2", "external-id": "nsx-vlan-transportzone-794", "segmentation_id": 794, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3515dec2-bd", "ovs_interfaceid": "3515dec2-bd6c-4f96-aff1-8dd32de6f7d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.385453] env[62460]: DEBUG oslo_vmware.api [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313565, 'name': PowerOnVM_Task, 'duration_secs': 0.611904} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.385453] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 829.385453] env[62460]: INFO nova.compute.manager [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] Took 10.44 seconds to spawn the instance on the hypervisor. [ 829.385453] env[62460]: DEBUG nova.compute.manager [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 829.385453] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a90de33d-b6b1-48c8-9731-9bf64f33574b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.481929] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7b25a346-a87d-486c-86ee-f9dc9c62cc1e tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.875s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.487143] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.298s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.487143] env[62460]: INFO nova.compute.claims [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 829.515018] env[62460]: INFO nova.scheduler.client.report [None req-7b25a346-a87d-486c-86ee-f9dc9c62cc1e tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Deleted allocations for instance 51718896-f5bf-43a9-9396-1ac768737ba2 [ 829.521044] env[62460]: INFO nova.compute.manager [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Took 33.86 seconds to build instance. [ 829.528460] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313568, 'name': CreateVM_Task, 'duration_secs': 0.593712} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.528562] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 829.529217] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.529385] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.529712] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 829.529969] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a89340a9-f075-4a03-93fd-ea3881369df8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.535529] env[62460]: DEBUG oslo_vmware.api [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Waiting for the task: (returnval){ [ 829.535529] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]5228c013-4740-6810-2138-92dd5404100d" [ 829.535529] env[62460]: _type = "Task" [ 829.535529] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.545333] env[62460]: DEBUG oslo_vmware.api [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5228c013-4740-6810-2138-92dd5404100d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.563957] env[62460]: DEBUG oslo_vmware.api [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313569, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.606239} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.564222] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] b76f37a0-91d0-4a01-9d95-9c6586081175/b76f37a0-91d0-4a01-9d95-9c6586081175.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 829.564443] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 829.564693] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-840379cd-1c63-4041-bbd5-efda353136f6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.572540] env[62460]: DEBUG oslo_vmware.api [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 829.572540] env[62460]: value = "task-1313570" [ 829.572540] env[62460]: _type = "Task" [ 829.572540] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.582254] env[62460]: DEBUG oslo_vmware.api [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313570, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.629344] env[62460]: DEBUG nova.compute.manager [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 829.654995] env[62460]: DEBUG nova.virt.hardware [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 829.655279] env[62460]: DEBUG nova.virt.hardware [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 829.655447] env[62460]: DEBUG nova.virt.hardware [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 829.655633] env[62460]: DEBUG nova.virt.hardware [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 829.655781] env[62460]: DEBUG nova.virt.hardware [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 829.655926] env[62460]: DEBUG nova.virt.hardware [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 829.656145] env[62460]: DEBUG nova.virt.hardware [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 829.656341] env[62460]: DEBUG nova.virt.hardware [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 829.656525] env[62460]: DEBUG nova.virt.hardware [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 829.656694] env[62460]: DEBUG nova.virt.hardware [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 829.656865] env[62460]: DEBUG nova.virt.hardware [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 829.657730] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1eaa886-ee87-4238-8bd4-58a734ae65ca {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.666131] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c733a1f4-d182-41d5-9cc1-b684bbc63252 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.723372] env[62460]: DEBUG oslo_concurrency.lockutils [req-419bee05-82c2-4230-b431-6607905232a9 req-c5b3fed4-5234-4f89-a7b7-393180fc054c service nova] Releasing lock "refresh_cache-5b6e8205-003e-49c0-a73d-be2e032a8272" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.902432] env[62460]: INFO nova.compute.manager [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] Took 37.60 seconds to build instance. [ 830.023648] env[62460]: DEBUG oslo_concurrency.lockutils [None req-71d93765-4afc-4bde-be9a-80acb38cc8c3 tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Lock "13890351-6091-4b4f-8484-1cdd0c8523b0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 108.550s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.027031] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7b25a346-a87d-486c-86ee-f9dc9c62cc1e tempest-ServerShowV257Test-1798436402 tempest-ServerShowV257Test-1798436402-project-member] Lock "51718896-f5bf-43a9-9396-1ac768737ba2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.429s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.050102] env[62460]: DEBUG oslo_vmware.api [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5228c013-4740-6810-2138-92dd5404100d, 'name': SearchDatastore_Task, 'duration_secs': 0.026107} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.050495] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.051382] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 830.051714] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.051901] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.052144] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 830.052861] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-edb00748-be3b-44cf-9ddd-dba18ab367ef {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.069306] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 830.069469] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 830.070218] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fb00c3ac-b9a3-4a58-97df-024a7f5be0aa {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.079447] env[62460]: DEBUG oslo_vmware.api [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Waiting for the task: (returnval){ [ 830.079447] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52c7c1e2-44eb-e4db-ef2c-a7a95348997b" [ 830.079447] env[62460]: _type = "Task" [ 830.079447] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.082780] env[62460]: DEBUG oslo_vmware.api [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313570, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066737} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.086233] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 830.086654] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c2624ccd-5dd5-44c5-9618-3e97c3395a8c tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Acquiring lock "13890351-6091-4b4f-8484-1cdd0c8523b0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.086934] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c2624ccd-5dd5-44c5-9618-3e97c3395a8c tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Lock "13890351-6091-4b4f-8484-1cdd0c8523b0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.087236] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c2624ccd-5dd5-44c5-9618-3e97c3395a8c tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Acquiring lock "13890351-6091-4b4f-8484-1cdd0c8523b0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.087943] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c2624ccd-5dd5-44c5-9618-3e97c3395a8c tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Lock "13890351-6091-4b4f-8484-1cdd0c8523b0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.088201] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c2624ccd-5dd5-44c5-9618-3e97c3395a8c tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Lock "13890351-6091-4b4f-8484-1cdd0c8523b0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.093163] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ba088e9-b042-48bb-b73b-500b287ab61c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.099174] env[62460]: INFO nova.compute.manager [None req-c2624ccd-5dd5-44c5-9618-3e97c3395a8c tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Terminating instance [ 830.099174] env[62460]: DEBUG nova.compute.manager [None req-c2624ccd-5dd5-44c5-9618-3e97c3395a8c tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 830.099174] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-c2624ccd-5dd5-44c5-9618-3e97c3395a8c tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 830.099735] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29fa2916-922c-4ed6-8a8e-ebc2c8eb6711 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.125672] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Reconfiguring VM instance instance-00000041 to attach disk [datastore2] b76f37a0-91d0-4a01-9d95-9c6586081175/b76f37a0-91d0-4a01-9d95-9c6586081175.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 830.126648] env[62460]: DEBUG oslo_vmware.api [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52c7c1e2-44eb-e4db-ef2c-a7a95348997b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.128958] env[62460]: DEBUG nova.compute.manager [req-e77593f6-56f5-4668-99ff-d93f10942298 req-a34ff66d-58b2-4726-860f-cea8fa0e770a service nova] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] Received event network-vif-plugged-608a7ef6-983b-4023-a6b6-69e632436554 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 830.129231] env[62460]: DEBUG oslo_concurrency.lockutils [req-e77593f6-56f5-4668-99ff-d93f10942298 req-a34ff66d-58b2-4726-860f-cea8fa0e770a service nova] Acquiring lock "9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.129492] env[62460]: DEBUG oslo_concurrency.lockutils [req-e77593f6-56f5-4668-99ff-d93f10942298 req-a34ff66d-58b2-4726-860f-cea8fa0e770a service nova] Lock "9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.129724] env[62460]: DEBUG oslo_concurrency.lockutils [req-e77593f6-56f5-4668-99ff-d93f10942298 req-a34ff66d-58b2-4726-860f-cea8fa0e770a service nova] Lock "9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.129956] env[62460]: DEBUG nova.compute.manager [req-e77593f6-56f5-4668-99ff-d93f10942298 req-a34ff66d-58b2-4726-860f-cea8fa0e770a service nova] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] No waiting events found dispatching network-vif-plugged-608a7ef6-983b-4023-a6b6-69e632436554 {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 830.130183] env[62460]: WARNING nova.compute.manager [req-e77593f6-56f5-4668-99ff-d93f10942298 req-a34ff66d-58b2-4726-860f-cea8fa0e770a service nova] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] Received unexpected event network-vif-plugged-608a7ef6-983b-4023-a6b6-69e632436554 for instance with vm_state building and task_state spawning. [ 830.130767] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b14882e7-fbc7-45ba-b23f-ea78243d1572 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.166243] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2624ccd-5dd5-44c5-9618-3e97c3395a8c tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 830.168398] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-84c111a0-0ce8-4ace-ae1d-2bb16e764f8c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.170752] env[62460]: DEBUG oslo_vmware.api [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 830.170752] env[62460]: value = "task-1313571" [ 830.170752] env[62460]: _type = "Task" [ 830.170752] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.180738] env[62460]: DEBUG nova.network.neutron [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] Successfully updated port: 608a7ef6-983b-4023-a6b6-69e632436554 {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 830.181048] env[62460]: DEBUG oslo_vmware.api [None req-c2624ccd-5dd5-44c5-9618-3e97c3395a8c tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Waiting for the task: (returnval){ [ 830.181048] env[62460]: value = "task-1313572" [ 830.181048] env[62460]: _type = "Task" [ 830.181048] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.191556] env[62460]: DEBUG oslo_vmware.api [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313571, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.199504] env[62460]: DEBUG oslo_vmware.api [None req-c2624ccd-5dd5-44c5-9618-3e97c3395a8c tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Task: {'id': task-1313572, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.405764] env[62460]: DEBUG oslo_concurrency.lockutils [None req-054ebb1d-c31b-4f23-ac71-f4e7f6c7271e tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Lock "fde12685-d3b9-46a0-8931-25b904d4f21e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 115.784s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.528763] env[62460]: DEBUG nova.compute.manager [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 830.596341] env[62460]: DEBUG oslo_vmware.api [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52c7c1e2-44eb-e4db-ef2c-a7a95348997b, 'name': SearchDatastore_Task, 'duration_secs': 0.059095} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.597173] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0694b281-982a-434b-b131-9ece2539912d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.607422] env[62460]: DEBUG oslo_vmware.api [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Waiting for the task: (returnval){ [ 830.607422] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]5276adf8-2b2e-0a5a-9de4-374878081dc0" [ 830.607422] env[62460]: _type = "Task" [ 830.607422] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.615955] env[62460]: DEBUG oslo_vmware.api [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5276adf8-2b2e-0a5a-9de4-374878081dc0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.686579] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Acquiring lock "refresh_cache-9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.686746] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Acquired lock "refresh_cache-9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.686869] env[62460]: DEBUG nova.network.neutron [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 830.688663] env[62460]: DEBUG oslo_vmware.api [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313571, 'name': ReconfigVM_Task, 'duration_secs': 0.478362} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.695547] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Reconfigured VM instance instance-00000041 to attach disk [datastore2] b76f37a0-91d0-4a01-9d95-9c6586081175/b76f37a0-91d0-4a01-9d95-9c6586081175.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 830.697175] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-35aaac13-48d4-4f7f-9955-aec153cdaa69 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.706024] env[62460]: DEBUG oslo_vmware.api [None req-c2624ccd-5dd5-44c5-9618-3e97c3395a8c tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Task: {'id': task-1313572, 'name': PowerOffVM_Task, 'duration_secs': 0.201721} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.707267] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2624ccd-5dd5-44c5-9618-3e97c3395a8c tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 830.707482] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-c2624ccd-5dd5-44c5-9618-3e97c3395a8c tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 830.707764] env[62460]: DEBUG oslo_vmware.api [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 830.707764] env[62460]: value = "task-1313573" [ 830.707764] env[62460]: _type = "Task" [ 830.707764] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.710263] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9b15e952-cfe8-48d8-8552-cd4c47b95f6c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.720636] env[62460]: DEBUG oslo_vmware.api [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313573, 'name': Rename_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.782773] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-c2624ccd-5dd5-44c5-9618-3e97c3395a8c tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 830.783014] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-c2624ccd-5dd5-44c5-9618-3e97c3395a8c tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Deleting contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 830.783194] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-c2624ccd-5dd5-44c5-9618-3e97c3395a8c tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Deleting the datastore file [datastore2] 13890351-6091-4b4f-8484-1cdd0c8523b0 {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 830.783510] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7b3fbc90-4301-4488-89f8-b6449de0214f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.791170] env[62460]: DEBUG oslo_vmware.api [None req-c2624ccd-5dd5-44c5-9618-3e97c3395a8c tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Waiting for the task: (returnval){ [ 830.791170] env[62460]: value = "task-1313575" [ 830.791170] env[62460]: _type = "Task" [ 830.791170] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.802797] env[62460]: DEBUG oslo_vmware.api [None req-c2624ccd-5dd5-44c5-9618-3e97c3395a8c tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Task: {'id': task-1313575, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.805826] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94b85a7b-59e6-4c4a-89be-fdc0a4ef1253 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.816032] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-260026d0-46e8-4f72-a66c-f85325fbee3c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.849907] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f34447a7-41dc-429c-a537-a70fc0633a17 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.858652] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d2466e6-1614-4314-9326-a1432af4b3c5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.873140] env[62460]: DEBUG nova.compute.provider_tree [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 830.908913] env[62460]: DEBUG nova.compute.manager [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 831.050124] env[62460]: DEBUG oslo_concurrency.lockutils [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.120615] env[62460]: DEBUG oslo_vmware.api [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5276adf8-2b2e-0a5a-9de4-374878081dc0, 'name': SearchDatastore_Task, 'duration_secs': 0.0114} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.120818] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.121086] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 5b6e8205-003e-49c0-a73d-be2e032a8272/5b6e8205-003e-49c0-a73d-be2e032a8272.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 831.123031] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2ba81e5a-2eb9-4f62-a91c-e44fad7d03a0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.131124] env[62460]: DEBUG oslo_vmware.api [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Waiting for the task: (returnval){ [ 831.131124] env[62460]: value = "task-1313576" [ 831.131124] env[62460]: _type = "Task" [ 831.131124] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.142568] env[62460]: DEBUG oslo_vmware.api [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1313576, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.225661] env[62460]: DEBUG oslo_vmware.api [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313573, 'name': Rename_Task, 'duration_secs': 0.163501} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.225661] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 831.225661] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9be64bc8-011e-4168-a8e7-fafaebe5d459 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.226216] env[62460]: DEBUG nova.network.neutron [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 831.234417] env[62460]: DEBUG oslo_vmware.api [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 831.234417] env[62460]: value = "task-1313577" [ 831.234417] env[62460]: _type = "Task" [ 831.234417] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.243633] env[62460]: DEBUG oslo_vmware.api [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313577, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.304417] env[62460]: DEBUG oslo_vmware.api [None req-c2624ccd-5dd5-44c5-9618-3e97c3395a8c tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Task: {'id': task-1313575, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.136291} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.304681] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-c2624ccd-5dd5-44c5-9618-3e97c3395a8c tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 831.304869] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-c2624ccd-5dd5-44c5-9618-3e97c3395a8c tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Deleted contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 831.305064] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-c2624ccd-5dd5-44c5-9618-3e97c3395a8c tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 831.305286] env[62460]: INFO nova.compute.manager [None req-c2624ccd-5dd5-44c5-9618-3e97c3395a8c tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Took 1.21 seconds to destroy the instance on the hypervisor. [ 831.305601] env[62460]: DEBUG oslo.service.loopingcall [None req-c2624ccd-5dd5-44c5-9618-3e97c3395a8c tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 831.305820] env[62460]: DEBUG nova.compute.manager [-] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 831.305988] env[62460]: DEBUG nova.network.neutron [-] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 831.376849] env[62460]: DEBUG nova.scheduler.client.report [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 831.433037] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.445319] env[62460]: DEBUG nova.network.neutron [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] Updating instance_info_cache with network_info: [{"id": "608a7ef6-983b-4023-a6b6-69e632436554", "address": "fa:16:3e:e5:9a:d5", "network": {"id": "1e15b3ec-bd84-41d8-92d0-66d17a474051", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-641826644-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "44220d464249478fb1373bc6b146ee0f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ce62383-8e84-4e26-955b-74c11392f4c9", "external-id": "nsx-vlan-transportzone-215", "segmentation_id": 215, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap608a7ef6-98", "ovs_interfaceid": "608a7ef6-983b-4023-a6b6-69e632436554", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.619028] env[62460]: DEBUG nova.compute.manager [req-8f894fb4-d16d-49b4-aabb-703b8175c2a6 req-2bf0876d-7df0-49c2-8dee-65b8075ff128 service nova] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Received event network-vif-deleted-c2c96f1c-79ab-4faa-a73b-9d11cae9a02f {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 831.619028] env[62460]: INFO nova.compute.manager [req-8f894fb4-d16d-49b4-aabb-703b8175c2a6 req-2bf0876d-7df0-49c2-8dee-65b8075ff128 service nova] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Neutron deleted interface c2c96f1c-79ab-4faa-a73b-9d11cae9a02f; detaching it from the instance and deleting it from the info cache [ 831.619028] env[62460]: DEBUG nova.network.neutron [req-8f894fb4-d16d-49b4-aabb-703b8175c2a6 req-2bf0876d-7df0-49c2-8dee-65b8075ff128 service nova] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.643579] env[62460]: DEBUG oslo_vmware.api [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1313576, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.746371] env[62460]: DEBUG oslo_vmware.api [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313577, 'name': PowerOnVM_Task} progress is 90%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.882161] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.398s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.882745] env[62460]: DEBUG nova.compute.manager [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 831.885825] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.483s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.887191] env[62460]: INFO nova.compute.claims [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 831.948041] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Releasing lock "refresh_cache-9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.948586] env[62460]: DEBUG nova.compute.manager [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] Instance network_info: |[{"id": "608a7ef6-983b-4023-a6b6-69e632436554", "address": "fa:16:3e:e5:9a:d5", "network": {"id": "1e15b3ec-bd84-41d8-92d0-66d17a474051", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-641826644-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "44220d464249478fb1373bc6b146ee0f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ce62383-8e84-4e26-955b-74c11392f4c9", "external-id": "nsx-vlan-transportzone-215", "segmentation_id": 215, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap608a7ef6-98", "ovs_interfaceid": "608a7ef6-983b-4023-a6b6-69e632436554", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 831.949170] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e5:9a:d5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2ce62383-8e84-4e26-955b-74c11392f4c9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '608a7ef6-983b-4023-a6b6-69e632436554', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 831.957239] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Creating folder: Project (44220d464249478fb1373bc6b146ee0f). Parent ref: group-v281134. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 831.957908] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-939b25c4-3a57-466e-a42e-b6ccbaa99377 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.976970] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Created folder: Project (44220d464249478fb1373bc6b146ee0f) in parent group-v281134. [ 831.976970] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Creating folder: Instances. Parent ref: group-v281186. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 831.977327] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cccb12e8-62ec-4e44-9d2f-16fd63a994d2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.987718] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Created folder: Instances in parent group-v281186. [ 831.987972] env[62460]: DEBUG oslo.service.loopingcall [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 831.988190] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 831.988530] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c00ee9fa-5597-4ac7-9626-d5995d8cb0bb {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.010955] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 832.010955] env[62460]: value = "task-1313580" [ 832.010955] env[62460]: _type = "Task" [ 832.010955] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.019446] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313580, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.092492] env[62460]: DEBUG nova.network.neutron [-] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.126769] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9a2d13df-c4b2-4c24-898e-d833e1f8cb8d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.141744] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adcc2dfc-f947-40bc-b4ca-ff72a5ecf4b9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.157172] env[62460]: DEBUG oslo_vmware.api [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1313576, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.585005} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.157877] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 5b6e8205-003e-49c0-a73d-be2e032a8272/5b6e8205-003e-49c0-a73d-be2e032a8272.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 832.158186] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 832.158414] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c77e75fb-bc21-4c25-bc4e-c1e45c751a84 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.167045] env[62460]: DEBUG oslo_vmware.api [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Waiting for the task: (returnval){ [ 832.167045] env[62460]: value = "task-1313581" [ 832.167045] env[62460]: _type = "Task" [ 832.167045] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.180242] env[62460]: DEBUG nova.compute.manager [req-8f894fb4-d16d-49b4-aabb-703b8175c2a6 req-2bf0876d-7df0-49c2-8dee-65b8075ff128 service nova] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Detach interface failed, port_id=c2c96f1c-79ab-4faa-a73b-9d11cae9a02f, reason: Instance 13890351-6091-4b4f-8484-1cdd0c8523b0 could not be found. {{(pid=62460) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 832.186545] env[62460]: DEBUG oslo_vmware.api [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1313581, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.233560] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3a5bc18a-5fa3-456c-bbd0-48ae5938b44c tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Acquiring lock "72e91f8e-0619-464c-b9bc-d6a14be42cb8" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.233947] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3a5bc18a-5fa3-456c-bbd0-48ae5938b44c tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Lock "72e91f8e-0619-464c-b9bc-d6a14be42cb8" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.234154] env[62460]: DEBUG nova.compute.manager [None req-3a5bc18a-5fa3-456c-bbd0-48ae5938b44c tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 832.235103] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9087e23b-5a82-4601-a2ad-a00733353c46 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.249320] env[62460]: DEBUG nova.compute.manager [None req-3a5bc18a-5fa3-456c-bbd0-48ae5938b44c tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62460) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 832.249858] env[62460]: DEBUG nova.objects.instance [None req-3a5bc18a-5fa3-456c-bbd0-48ae5938b44c tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Lazy-loading 'flavor' on Instance uuid 72e91f8e-0619-464c-b9bc-d6a14be42cb8 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 832.251751] env[62460]: DEBUG oslo_vmware.api [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313577, 'name': PowerOnVM_Task, 'duration_secs': 0.575366} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.252145] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 832.252186] env[62460]: INFO nova.compute.manager [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Took 8.50 seconds to spawn the instance on the hypervisor. [ 832.252518] env[62460]: DEBUG nova.compute.manager [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 832.253457] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4335f094-750c-4b32-b7e5-915616c68d2e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.258068] env[62460]: DEBUG nova.compute.manager [req-961d0ea9-c14f-4a66-b36b-ee29cc52c7e2 req-f75e8c28-0847-4d09-adef-6a78582d844d service nova] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] Received event network-changed-608a7ef6-983b-4023-a6b6-69e632436554 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 832.258257] env[62460]: DEBUG nova.compute.manager [req-961d0ea9-c14f-4a66-b36b-ee29cc52c7e2 req-f75e8c28-0847-4d09-adef-6a78582d844d service nova] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] Refreshing instance network info cache due to event network-changed-608a7ef6-983b-4023-a6b6-69e632436554. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 832.258461] env[62460]: DEBUG oslo_concurrency.lockutils [req-961d0ea9-c14f-4a66-b36b-ee29cc52c7e2 req-f75e8c28-0847-4d09-adef-6a78582d844d service nova] Acquiring lock "refresh_cache-9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.258607] env[62460]: DEBUG oslo_concurrency.lockutils [req-961d0ea9-c14f-4a66-b36b-ee29cc52c7e2 req-f75e8c28-0847-4d09-adef-6a78582d844d service nova] Acquired lock "refresh_cache-9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.258765] env[62460]: DEBUG nova.network.neutron [req-961d0ea9-c14f-4a66-b36b-ee29cc52c7e2 req-f75e8c28-0847-4d09-adef-6a78582d844d service nova] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] Refreshing network info cache for port 608a7ef6-983b-4023-a6b6-69e632436554 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 832.391661] env[62460]: DEBUG nova.compute.utils [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 832.395083] env[62460]: DEBUG nova.compute.manager [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 832.395341] env[62460]: DEBUG nova.network.neutron [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 832.450868] env[62460]: DEBUG nova.policy [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b4c1b6d1942542a7815aee4ec860259a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '44220d464249478fb1373bc6b146ee0f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 832.524227] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313580, 'name': CreateVM_Task} progress is 99%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.595399] env[62460]: INFO nova.compute.manager [-] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Took 1.29 seconds to deallocate network for instance. [ 832.678400] env[62460]: DEBUG oslo_vmware.api [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1313581, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.090543} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.678797] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 832.679891] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d77dc9d9-579a-45f4-a67c-48e4e5956140 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.706216] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Reconfiguring VM instance instance-00000042 to attach disk [datastore1] 5b6e8205-003e-49c0-a73d-be2e032a8272/5b6e8205-003e-49c0-a73d-be2e032a8272.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 832.706482] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-19065524-96aa-4691-aecb-4c6b38661aed {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.728178] env[62460]: DEBUG oslo_vmware.api [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Waiting for the task: (returnval){ [ 832.728178] env[62460]: value = "task-1313582" [ 832.728178] env[62460]: _type = "Task" [ 832.728178] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.738095] env[62460]: DEBUG oslo_vmware.api [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1313582, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.749545] env[62460]: DEBUG nova.network.neutron [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] Successfully created port: 42bbd4ac-f793-4c77-96c8-d45f34b8bb15 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 832.755539] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a5bc18a-5fa3-456c-bbd0-48ae5938b44c tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 832.755801] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ce1e8187-eddd-4cbf-adda-1ac2a15a0260 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.765681] env[62460]: DEBUG oslo_vmware.api [None req-3a5bc18a-5fa3-456c-bbd0-48ae5938b44c tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Waiting for the task: (returnval){ [ 832.765681] env[62460]: value = "task-1313583" [ 832.765681] env[62460]: _type = "Task" [ 832.765681] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.779829] env[62460]: DEBUG oslo_vmware.api [None req-3a5bc18a-5fa3-456c-bbd0-48ae5938b44c tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313583, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.783431] env[62460]: INFO nova.compute.manager [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Took 33.69 seconds to build instance. [ 832.895653] env[62460]: DEBUG nova.compute.manager [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 833.009929] env[62460]: DEBUG nova.network.neutron [req-961d0ea9-c14f-4a66-b36b-ee29cc52c7e2 req-f75e8c28-0847-4d09-adef-6a78582d844d service nova] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] Updated VIF entry in instance network info cache for port 608a7ef6-983b-4023-a6b6-69e632436554. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 833.010428] env[62460]: DEBUG nova.network.neutron [req-961d0ea9-c14f-4a66-b36b-ee29cc52c7e2 req-f75e8c28-0847-4d09-adef-6a78582d844d service nova] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] Updating instance_info_cache with network_info: [{"id": "608a7ef6-983b-4023-a6b6-69e632436554", "address": "fa:16:3e:e5:9a:d5", "network": {"id": "1e15b3ec-bd84-41d8-92d0-66d17a474051", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-641826644-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "44220d464249478fb1373bc6b146ee0f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ce62383-8e84-4e26-955b-74c11392f4c9", "external-id": "nsx-vlan-transportzone-215", "segmentation_id": 215, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap608a7ef6-98", "ovs_interfaceid": "608a7ef6-983b-4023-a6b6-69e632436554", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.025539] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313580, 'name': CreateVM_Task, 'duration_secs': 0.515587} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.025810] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 833.026555] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.026789] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.027176] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 833.027845] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bb32c023-c1fc-474c-8edd-307011d6b901 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.036540] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Waiting for the task: (returnval){ [ 833.036540] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]526481c7-c84a-b22c-c095-807879d73edc" [ 833.036540] env[62460]: _type = "Task" [ 833.036540] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.050254] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]526481c7-c84a-b22c-c095-807879d73edc, 'name': SearchDatastore_Task, 'duration_secs': 0.010494} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.050640] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.050928] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 833.051252] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.051456] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.051690] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 833.051999] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-92e8e2f1-08e6-4e96-b21e-b8a90ae0db48 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.061810] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 833.062032] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 833.062777] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e297e65b-d612-4148-a2c9-b56bc325067a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.074701] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Waiting for the task: (returnval){ [ 833.074701] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]5238a882-e01c-dab4-8805-7e2d90635667" [ 833.074701] env[62460]: _type = "Task" [ 833.074701] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.085869] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5238a882-e01c-dab4-8805-7e2d90635667, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.103890] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c2624ccd-5dd5-44c5-9618-3e97c3395a8c tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.237104] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61d51a39-3268-484d-ab93-39ed23f531a6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.245830] env[62460]: DEBUG oslo_vmware.api [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1313582, 'name': ReconfigVM_Task, 'duration_secs': 0.30768} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.246624] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Reconfigured VM instance instance-00000042 to attach disk [datastore1] 5b6e8205-003e-49c0-a73d-be2e032a8272/5b6e8205-003e-49c0-a73d-be2e032a8272.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 833.247332] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-242edf9c-69d5-4064-beb0-89327be0d46a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.252514] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-059cb126-675e-400d-96b4-8cef9a11b3a2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.291797] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d8835c53-f907-4c79-92ca-e5ba3480e88c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "b76f37a0-91d0-4a01-9d95-9c6586081175" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 103.879s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.292772] env[62460]: DEBUG oslo_vmware.api [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Waiting for the task: (returnval){ [ 833.292772] env[62460]: value = "task-1313584" [ 833.292772] env[62460]: _type = "Task" [ 833.292772] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.298786] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b1a3bf2-202b-4a6b-94a9-817c06015396 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.318555] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd072d1b-10c8-43c5-8a39-899e54eed709 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.322913] env[62460]: DEBUG oslo_vmware.api [None req-3a5bc18a-5fa3-456c-bbd0-48ae5938b44c tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313583, 'name': PowerOffVM_Task, 'duration_secs': 0.379531} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.323129] env[62460]: DEBUG oslo_vmware.api [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1313584, 'name': Rename_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.323417] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a5bc18a-5fa3-456c-bbd0-48ae5938b44c tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 833.323667] env[62460]: DEBUG nova.compute.manager [None req-3a5bc18a-5fa3-456c-bbd0-48ae5938b44c tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 833.324792] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d159aa9c-3fc5-4866-ad86-fb4f0a97e77f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.339707] env[62460]: DEBUG nova.compute.provider_tree [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 833.515958] env[62460]: DEBUG oslo_concurrency.lockutils [req-961d0ea9-c14f-4a66-b36b-ee29cc52c7e2 req-f75e8c28-0847-4d09-adef-6a78582d844d service nova] Releasing lock "refresh_cache-9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.586951] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5238a882-e01c-dab4-8805-7e2d90635667, 'name': SearchDatastore_Task, 'duration_secs': 0.027158} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.587825] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f87119c2-a4b4-4924-8ff4-4d2f8c3d798f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.593062] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Waiting for the task: (returnval){ [ 833.593062] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]5244ed0d-ec99-b345-535a-5fb5bd880d83" [ 833.593062] env[62460]: _type = "Task" [ 833.593062] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.601452] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5244ed0d-ec99-b345-535a-5fb5bd880d83, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.803237] env[62460]: DEBUG nova.compute.manager [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 833.817885] env[62460]: DEBUG oslo_vmware.api [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1313584, 'name': Rename_Task, 'duration_secs': 0.172449} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.818190] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 833.818573] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1d8c0c1f-b376-44a6-87db-11b1153f2e3b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.827525] env[62460]: DEBUG oslo_vmware.api [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Waiting for the task: (returnval){ [ 833.827525] env[62460]: value = "task-1313585" [ 833.827525] env[62460]: _type = "Task" [ 833.827525] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.850068] env[62460]: DEBUG nova.scheduler.client.report [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 833.857789] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3a5bc18a-5fa3-456c-bbd0-48ae5938b44c tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Lock "72e91f8e-0619-464c-b9bc-d6a14be42cb8" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.624s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.912477] env[62460]: DEBUG nova.compute.manager [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 833.934518] env[62460]: DEBUG nova.virt.hardware [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 833.934780] env[62460]: DEBUG nova.virt.hardware [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 833.934935] env[62460]: DEBUG nova.virt.hardware [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 833.935133] env[62460]: DEBUG nova.virt.hardware [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 833.935292] env[62460]: DEBUG nova.virt.hardware [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 833.935466] env[62460]: DEBUG nova.virt.hardware [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 833.935684] env[62460]: DEBUG nova.virt.hardware [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 833.935850] env[62460]: DEBUG nova.virt.hardware [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 833.936029] env[62460]: DEBUG nova.virt.hardware [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 833.936201] env[62460]: DEBUG nova.virt.hardware [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 833.936410] env[62460]: DEBUG nova.virt.hardware [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 833.937592] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e8ce98c-a781-411e-b4e1-8d4b550946cc {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.946666] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eba461c2-992a-424c-b926-0cb84a8618f3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.104181] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5244ed0d-ec99-b345-535a-5fb5bd880d83, 'name': SearchDatastore_Task, 'duration_secs': 0.009905} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.104551] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.104845] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b/9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 834.105127] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-30b3f18a-7d75-46b8-b2f5-76be6017e2bf {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.114565] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Waiting for the task: (returnval){ [ 834.114565] env[62460]: value = "task-1313586" [ 834.114565] env[62460]: _type = "Task" [ 834.114565] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.123332] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313586, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.127861] env[62460]: INFO nova.compute.manager [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Rebuilding instance [ 834.170780] env[62460]: DEBUG nova.compute.manager [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 834.171686] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-086045d0-ea51-4fc9-9535-b0970a9a4bbf {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.333802] env[62460]: DEBUG oslo_concurrency.lockutils [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.339981] env[62460]: DEBUG oslo_vmware.api [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1313585, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.356654] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.471s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.357241] env[62460]: DEBUG nova.compute.manager [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 834.360140] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9c83d044-4a1e-4d35-a1ca-cefdd159ac66 tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.806s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.360293] env[62460]: DEBUG nova.objects.instance [None req-9c83d044-4a1e-4d35-a1ca-cefdd159ac66 tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Lazy-loading 'resources' on Instance uuid db09d1f5-88cc-4dc7-9a7b-5d53d09567fa {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 834.627648] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313586, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.657493] env[62460]: DEBUG nova.compute.manager [req-dee0deea-386a-44cf-859e-0b12d877646d req-f09666ec-b1e6-4eeb-970a-27e13b156a95 service nova] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] Received event network-vif-plugged-42bbd4ac-f793-4c77-96c8-d45f34b8bb15 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 834.657722] env[62460]: DEBUG oslo_concurrency.lockutils [req-dee0deea-386a-44cf-859e-0b12d877646d req-f09666ec-b1e6-4eeb-970a-27e13b156a95 service nova] Acquiring lock "0269dc64-d2b1-43c5-bdf7-11d97e534819-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.657941] env[62460]: DEBUG oslo_concurrency.lockutils [req-dee0deea-386a-44cf-859e-0b12d877646d req-f09666ec-b1e6-4eeb-970a-27e13b156a95 service nova] Lock "0269dc64-d2b1-43c5-bdf7-11d97e534819-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.658145] env[62460]: DEBUG oslo_concurrency.lockutils [req-dee0deea-386a-44cf-859e-0b12d877646d req-f09666ec-b1e6-4eeb-970a-27e13b156a95 service nova] Lock "0269dc64-d2b1-43c5-bdf7-11d97e534819-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.658378] env[62460]: DEBUG nova.compute.manager [req-dee0deea-386a-44cf-859e-0b12d877646d req-f09666ec-b1e6-4eeb-970a-27e13b156a95 service nova] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] No waiting events found dispatching network-vif-plugged-42bbd4ac-f793-4c77-96c8-d45f34b8bb15 {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 834.658580] env[62460]: WARNING nova.compute.manager [req-dee0deea-386a-44cf-859e-0b12d877646d req-f09666ec-b1e6-4eeb-970a-27e13b156a95 service nova] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] Received unexpected event network-vif-plugged-42bbd4ac-f793-4c77-96c8-d45f34b8bb15 for instance with vm_state building and task_state spawning. [ 834.673465] env[62460]: DEBUG nova.network.neutron [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] Successfully updated port: 42bbd4ac-f793-4c77-96c8-d45f34b8bb15 {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 834.683482] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 834.683773] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-266803fe-d3b4-41d5-acd2-e62b86569272 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.696027] env[62460]: DEBUG oslo_vmware.api [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 834.696027] env[62460]: value = "task-1313587" [ 834.696027] env[62460]: _type = "Task" [ 834.696027] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.706127] env[62460]: DEBUG oslo_vmware.api [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313587, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.838909] env[62460]: DEBUG oslo_vmware.api [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1313585, 'name': PowerOnVM_Task, 'duration_secs': 0.534119} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.839190] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 834.839400] env[62460]: INFO nova.compute.manager [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Took 8.71 seconds to spawn the instance on the hypervisor. [ 834.839587] env[62460]: DEBUG nova.compute.manager [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 834.840389] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62748371-c6a5-4bdc-92f7-0e191de373d7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.866888] env[62460]: DEBUG nova.compute.utils [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 834.868816] env[62460]: DEBUG nova.compute.manager [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Not allocating networking since 'none' was specified. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 835.108358] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a3a6535-a97f-4d73-af1f-c8db443187af {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.119396] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04ebb5e5-ece6-4486-98b9-86856b2fc2bf {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.128510] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313586, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.615703} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.154032] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b/9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 835.154492] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 835.155233] env[62460]: DEBUG nova.objects.instance [None req-cd35464f-2a1a-4ba3-9e7f-0a7ed43a0290 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Lazy-loading 'flavor' on Instance uuid 72e91f8e-0619-464c-b9bc-d6a14be42cb8 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 835.157529] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-678d4f41-843b-433b-b659-8e8b23e97b4c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.160237] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a42ecc95-6e81-4cfe-bc12-df446fa3caa9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.170437] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be8b40bb-5219-4009-82d4-9f3229862292 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.175740] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Waiting for the task: (returnval){ [ 835.175740] env[62460]: value = "task-1313588" [ 835.175740] env[62460]: _type = "Task" [ 835.175740] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.176953] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Acquiring lock "refresh_cache-0269dc64-d2b1-43c5-bdf7-11d97e534819" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.177121] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Acquired lock "refresh_cache-0269dc64-d2b1-43c5-bdf7-11d97e534819" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.177243] env[62460]: DEBUG nova.network.neutron [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 835.192295] env[62460]: DEBUG nova.compute.provider_tree [None req-9c83d044-4a1e-4d35-a1ca-cefdd159ac66 tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 835.202170] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313588, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.207861] env[62460]: DEBUG oslo_vmware.api [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313587, 'name': PowerOffVM_Task, 'duration_secs': 0.274457} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.208143] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 835.208363] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 835.209136] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99348623-9d05-4efa-8ae8-2c535a9e6c23 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.216802] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 835.217048] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-34ff7908-81ae-4d82-a645-91c8c13f9aaf {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.296524] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 835.296770] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Deleting contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 835.296955] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Deleting the datastore file [datastore2] b76f37a0-91d0-4a01-9d95-9c6586081175 {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 835.297246] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-668839ae-5be6-4396-8687-33b338596232 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.303920] env[62460]: DEBUG oslo_vmware.api [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 835.303920] env[62460]: value = "task-1313590" [ 835.303920] env[62460]: _type = "Task" [ 835.303920] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.312371] env[62460]: DEBUG oslo_vmware.api [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313590, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.359908] env[62460]: INFO nova.compute.manager [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Took 33.56 seconds to build instance. [ 835.370627] env[62460]: DEBUG nova.compute.manager [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 835.662978] env[62460]: DEBUG oslo_concurrency.lockutils [None req-cd35464f-2a1a-4ba3-9e7f-0a7ed43a0290 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Acquiring lock "refresh_cache-72e91f8e-0619-464c-b9bc-d6a14be42cb8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.663292] env[62460]: DEBUG oslo_concurrency.lockutils [None req-cd35464f-2a1a-4ba3-9e7f-0a7ed43a0290 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Acquired lock "refresh_cache-72e91f8e-0619-464c-b9bc-d6a14be42cb8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.663573] env[62460]: DEBUG nova.network.neutron [None req-cd35464f-2a1a-4ba3-9e7f-0a7ed43a0290 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 835.663838] env[62460]: DEBUG nova.objects.instance [None req-cd35464f-2a1a-4ba3-9e7f-0a7ed43a0290 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Lazy-loading 'info_cache' on Instance uuid 72e91f8e-0619-464c-b9bc-d6a14be42cb8 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 835.690560] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313588, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.411702} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.691836] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 835.692906] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94a43941-2402-4f75-932f-3103a66c99d5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.719212] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] Reconfiguring VM instance instance-00000043 to attach disk [datastore1] 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b/9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 835.720075] env[62460]: ERROR nova.scheduler.client.report [None req-9c83d044-4a1e-4d35-a1ca-cefdd159ac66 tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] [req-9e521a52-cc14-4bb8-bb38-c163af9b3509] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 76e8f00d-65cc-4766-bb9d-5006fb51c728. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-9e521a52-cc14-4bb8-bb38-c163af9b3509"}]} [ 835.720679] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f66c6f16-585c-4d8a-bd2d-9c47e01d8d20 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.736961] env[62460]: DEBUG nova.network.neutron [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 835.745773] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Waiting for the task: (returnval){ [ 835.745773] env[62460]: value = "task-1313591" [ 835.745773] env[62460]: _type = "Task" [ 835.745773] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.755729] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313591, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.758376] env[62460]: DEBUG nova.scheduler.client.report [None req-9c83d044-4a1e-4d35-a1ca-cefdd159ac66 tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Refreshing inventories for resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 835.773584] env[62460]: DEBUG nova.scheduler.client.report [None req-9c83d044-4a1e-4d35-a1ca-cefdd159ac66 tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Updating ProviderTree inventory for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 835.773810] env[62460]: DEBUG nova.compute.provider_tree [None req-9c83d044-4a1e-4d35-a1ca-cefdd159ac66 tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 835.784818] env[62460]: DEBUG nova.scheduler.client.report [None req-9c83d044-4a1e-4d35-a1ca-cefdd159ac66 tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Refreshing aggregate associations for resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728, aggregates: None {{(pid=62460) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 835.802473] env[62460]: DEBUG nova.scheduler.client.report [None req-9c83d044-4a1e-4d35-a1ca-cefdd159ac66 tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Refreshing trait associations for resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62460) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 835.814057] env[62460]: DEBUG oslo_vmware.api [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313590, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.862265] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b172b0c2-42a4-446b-a682-bd543175ab8a tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lock "5b6e8205-003e-49c0-a73d-be2e032a8272" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 104.609s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.945559] env[62460]: DEBUG nova.network.neutron [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] Updating instance_info_cache with network_info: [{"id": "42bbd4ac-f793-4c77-96c8-d45f34b8bb15", "address": "fa:16:3e:09:33:db", "network": {"id": "1e15b3ec-bd84-41d8-92d0-66d17a474051", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-641826644-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "44220d464249478fb1373bc6b146ee0f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ce62383-8e84-4e26-955b-74c11392f4c9", "external-id": "nsx-vlan-transportzone-215", "segmentation_id": 215, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap42bbd4ac-f7", "ovs_interfaceid": "42bbd4ac-f793-4c77-96c8-d45f34b8bb15", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.079084] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b05d587-e308-4590-8451-8db1f368d55d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.087186] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8dc7478-6ade-4a6f-afdf-3078c29d7f8a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.117359] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d24d699-7793-42cb-a3c0-eefd6bf71119 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.125472] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0682924-8914-425b-97c8-d0a5d3745d1c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.141825] env[62460]: DEBUG nova.compute.provider_tree [None req-9c83d044-4a1e-4d35-a1ca-cefdd159ac66 tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 836.167476] env[62460]: DEBUG nova.objects.base [None req-cd35464f-2a1a-4ba3-9e7f-0a7ed43a0290 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Object Instance<72e91f8e-0619-464c-b9bc-d6a14be42cb8> lazy-loaded attributes: flavor,info_cache {{(pid=62460) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 836.256251] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313591, 'name': ReconfigVM_Task, 'duration_secs': 0.296729} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.256529] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] Reconfigured VM instance instance-00000043 to attach disk [datastore1] 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b/9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 836.257149] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7655ff90-09d0-467c-85ea-91f101702a0b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.264248] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Waiting for the task: (returnval){ [ 836.264248] env[62460]: value = "task-1313592" [ 836.264248] env[62460]: _type = "Task" [ 836.264248] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.273573] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313592, 'name': Rename_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.314332] env[62460]: DEBUG oslo_vmware.api [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313590, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.646536} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.314540] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 836.314743] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Deleted contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 836.314941] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 836.367553] env[62460]: DEBUG nova.compute.manager [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 836.381629] env[62460]: DEBUG nova.compute.manager [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 836.415829] env[62460]: DEBUG nova.virt.hardware [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 836.416101] env[62460]: DEBUG nova.virt.hardware [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 836.416263] env[62460]: DEBUG nova.virt.hardware [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 836.416473] env[62460]: DEBUG nova.virt.hardware [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 836.416632] env[62460]: DEBUG nova.virt.hardware [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 836.416785] env[62460]: DEBUG nova.virt.hardware [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 836.416995] env[62460]: DEBUG nova.virt.hardware [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 836.418351] env[62460]: DEBUG nova.virt.hardware [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 836.418548] env[62460]: DEBUG nova.virt.hardware [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 836.418724] env[62460]: DEBUG nova.virt.hardware [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 836.418906] env[62460]: DEBUG nova.virt.hardware [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 836.420381] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-230375e5-6b8f-4f65-b93d-2ab4f596305e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.429316] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-328e083e-a8f9-4d24-853f-c449509822de {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.443868] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Instance VIF info [] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 836.449376] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Creating folder: Project (266d65bb4ffb4bc1b7b675d61304912d). Parent ref: group-v281134. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 836.449881] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Releasing lock "refresh_cache-0269dc64-d2b1-43c5-bdf7-11d97e534819" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.450183] env[62460]: DEBUG nova.compute.manager [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] Instance network_info: |[{"id": "42bbd4ac-f793-4c77-96c8-d45f34b8bb15", "address": "fa:16:3e:09:33:db", "network": {"id": "1e15b3ec-bd84-41d8-92d0-66d17a474051", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-641826644-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "44220d464249478fb1373bc6b146ee0f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ce62383-8e84-4e26-955b-74c11392f4c9", "external-id": "nsx-vlan-transportzone-215", "segmentation_id": 215, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap42bbd4ac-f7", "ovs_interfaceid": "42bbd4ac-f793-4c77-96c8-d45f34b8bb15", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 836.450594] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5ff35f47-e364-488c-ac2f-03675f0a4de5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.452316] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:09:33:db', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2ce62383-8e84-4e26-955b-74c11392f4c9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '42bbd4ac-f793-4c77-96c8-d45f34b8bb15', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 836.459447] env[62460]: DEBUG oslo.service.loopingcall [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 836.459686] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 836.460591] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3fddb63a-636c-4da5-a3c3-6e9515ecbd2f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.481319] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Created folder: Project (266d65bb4ffb4bc1b7b675d61304912d) in parent group-v281134. [ 836.481490] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Creating folder: Instances. Parent ref: group-v281189. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 836.486191] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1a58829e-9c59-4993-86a6-3c3a521b9964 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.489040] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 836.489040] env[62460]: value = "task-1313594" [ 836.489040] env[62460]: _type = "Task" [ 836.489040] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.498383] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313594, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.499907] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Created folder: Instances in parent group-v281189. [ 836.500260] env[62460]: DEBUG oslo.service.loopingcall [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 836.500478] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 836.500685] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-400a5f52-6da5-4271-a3d7-44e3d64f799a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.518858] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 836.518858] env[62460]: value = "task-1313596" [ 836.518858] env[62460]: _type = "Task" [ 836.518858] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.527986] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313596, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.681032] env[62460]: DEBUG nova.scheduler.client.report [None req-9c83d044-4a1e-4d35-a1ca-cefdd159ac66 tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Updated inventory for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with generation 85 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 836.681328] env[62460]: DEBUG nova.compute.provider_tree [None req-9c83d044-4a1e-4d35-a1ca-cefdd159ac66 tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Updating resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 generation from 85 to 86 during operation: update_inventory {{(pid=62460) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 836.681584] env[62460]: DEBUG nova.compute.provider_tree [None req-9c83d044-4a1e-4d35-a1ca-cefdd159ac66 tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 836.765041] env[62460]: DEBUG nova.compute.manager [req-d835444e-5f10-494a-aa54-abcc2561dfd5 req-7e954fdc-9dc9-420b-ae32-85a5de614706 service nova] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] Received event network-changed-42bbd4ac-f793-4c77-96c8-d45f34b8bb15 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 836.765489] env[62460]: DEBUG nova.compute.manager [req-d835444e-5f10-494a-aa54-abcc2561dfd5 req-7e954fdc-9dc9-420b-ae32-85a5de614706 service nova] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] Refreshing instance network info cache due to event network-changed-42bbd4ac-f793-4c77-96c8-d45f34b8bb15. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 836.765787] env[62460]: DEBUG oslo_concurrency.lockutils [req-d835444e-5f10-494a-aa54-abcc2561dfd5 req-7e954fdc-9dc9-420b-ae32-85a5de614706 service nova] Acquiring lock "refresh_cache-0269dc64-d2b1-43c5-bdf7-11d97e534819" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.766060] env[62460]: DEBUG oslo_concurrency.lockutils [req-d835444e-5f10-494a-aa54-abcc2561dfd5 req-7e954fdc-9dc9-420b-ae32-85a5de614706 service nova] Acquired lock "refresh_cache-0269dc64-d2b1-43c5-bdf7-11d97e534819" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.766290] env[62460]: DEBUG nova.network.neutron [req-d835444e-5f10-494a-aa54-abcc2561dfd5 req-7e954fdc-9dc9-420b-ae32-85a5de614706 service nova] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] Refreshing network info cache for port 42bbd4ac-f793-4c77-96c8-d45f34b8bb15 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 836.782638] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313592, 'name': Rename_Task, 'duration_secs': 0.142132} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.783518] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 836.783769] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e6bc5feb-cf16-4486-876e-ca2421cb33b8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.795173] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Waiting for the task: (returnval){ [ 836.795173] env[62460]: value = "task-1313597" [ 836.795173] env[62460]: _type = "Task" [ 836.795173] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.807190] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313597, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.894279] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.972591] env[62460]: DEBUG nova.network.neutron [None req-cd35464f-2a1a-4ba3-9e7f-0a7ed43a0290 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Updating instance_info_cache with network_info: [{"id": "215f9da2-99ec-4af7-8488-b7185454b285", "address": "fa:16:3e:de:e8:6f", "network": {"id": "7428294f-e3f7-45f4-b85a-e51d132281d0", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-867124390-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f26e25609a64eca836f4d9dcdcfd666", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap215f9da2-99", "ovs_interfaceid": "215f9da2-99ec-4af7-8488-b7185454b285", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 837.000641] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313594, 'name': CreateVM_Task, 'duration_secs': 0.433268} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.000816] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 837.001519] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.001683] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.002120] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 837.002628] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-697e0cf0-d176-43cd-8071-a847270dbe56 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.008609] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Waiting for the task: (returnval){ [ 837.008609] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]5264e6fc-4513-eb82-63bc-e738079d963a" [ 837.008609] env[62460]: _type = "Task" [ 837.008609] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.019351] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5264e6fc-4513-eb82-63bc-e738079d963a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.028132] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313596, 'name': CreateVM_Task, 'duration_secs': 0.301219} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.028311] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 837.028717] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.028883] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.029258] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 837.029516] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ad5f697e-6a24-4f30-873c-7cde8b85e2d4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.035049] env[62460]: DEBUG oslo_vmware.api [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Waiting for the task: (returnval){ [ 837.035049] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52b1d499-6c26-b704-7116-bf9c1bf1eddc" [ 837.035049] env[62460]: _type = "Task" [ 837.035049] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.043523] env[62460]: DEBUG oslo_vmware.api [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52b1d499-6c26-b704-7116-bf9c1bf1eddc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.186643] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9c83d044-4a1e-4d35-a1ca-cefdd159ac66 tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.826s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.189389] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.033s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.190991] env[62460]: INFO nova.compute.claims [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 837.207561] env[62460]: INFO nova.scheduler.client.report [None req-9c83d044-4a1e-4d35-a1ca-cefdd159ac66 tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Deleted allocations for instance db09d1f5-88cc-4dc7-9a7b-5d53d09567fa [ 837.306251] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313597, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.351381] env[62460]: DEBUG nova.virt.hardware [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 837.351642] env[62460]: DEBUG nova.virt.hardware [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 837.351803] env[62460]: DEBUG nova.virt.hardware [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 837.351994] env[62460]: DEBUG nova.virt.hardware [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 837.352157] env[62460]: DEBUG nova.virt.hardware [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 837.352308] env[62460]: DEBUG nova.virt.hardware [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 837.352516] env[62460]: DEBUG nova.virt.hardware [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 837.352696] env[62460]: DEBUG nova.virt.hardware [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 837.352871] env[62460]: DEBUG nova.virt.hardware [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 837.353046] env[62460]: DEBUG nova.virt.hardware [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 837.353223] env[62460]: DEBUG nova.virt.hardware [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 837.354092] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-268316b8-7452-4f1f-a807-66a0656f564e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.364554] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e84f7281-c0b0-45fd-8827-734b80859d72 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.377974] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:25:53:33', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '04ccbc7a-cf8d-4ea2-8411-291a1e27df7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bcdde0bd-b227-4477-ae73-bebbae418c5e', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 837.385410] env[62460]: DEBUG oslo.service.loopingcall [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 837.387826] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 837.388062] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-67737380-29e6-4e5c-97d8-cab05b3778d1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.408950] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 837.408950] env[62460]: value = "task-1313598" [ 837.408950] env[62460]: _type = "Task" [ 837.408950] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.416801] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313598, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.475478] env[62460]: DEBUG oslo_concurrency.lockutils [None req-cd35464f-2a1a-4ba3-9e7f-0a7ed43a0290 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Releasing lock "refresh_cache-72e91f8e-0619-464c-b9bc-d6a14be42cb8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.520014] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5264e6fc-4513-eb82-63bc-e738079d963a, 'name': SearchDatastore_Task, 'duration_secs': 0.010759} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.520351] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.520588] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 837.520827] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.520976] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.521198] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 837.521477] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5c095356-4738-41d9-91e0-f3fd477cf295 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.532241] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 837.532449] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 837.533182] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6f736e97-42e1-4c41-ad75-f1a8d58e13f6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.536381] env[62460]: DEBUG nova.network.neutron [req-d835444e-5f10-494a-aa54-abcc2561dfd5 req-7e954fdc-9dc9-420b-ae32-85a5de614706 service nova] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] Updated VIF entry in instance network info cache for port 42bbd4ac-f793-4c77-96c8-d45f34b8bb15. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 837.536701] env[62460]: DEBUG nova.network.neutron [req-d835444e-5f10-494a-aa54-abcc2561dfd5 req-7e954fdc-9dc9-420b-ae32-85a5de614706 service nova] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] Updating instance_info_cache with network_info: [{"id": "42bbd4ac-f793-4c77-96c8-d45f34b8bb15", "address": "fa:16:3e:09:33:db", "network": {"id": "1e15b3ec-bd84-41d8-92d0-66d17a474051", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-641826644-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "44220d464249478fb1373bc6b146ee0f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ce62383-8e84-4e26-955b-74c11392f4c9", "external-id": "nsx-vlan-transportzone-215", "segmentation_id": 215, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap42bbd4ac-f7", "ovs_interfaceid": "42bbd4ac-f793-4c77-96c8-d45f34b8bb15", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 837.546363] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Waiting for the task: (returnval){ [ 837.546363] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52ca3126-e77a-d1ce-500d-dc3cb001d3e3" [ 837.546363] env[62460]: _type = "Task" [ 837.546363] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.548782] env[62460]: DEBUG oslo_vmware.api [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52b1d499-6c26-b704-7116-bf9c1bf1eddc, 'name': SearchDatastore_Task, 'duration_secs': 0.009293} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.554571] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.554810] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 837.555046] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.555204] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.555401] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 837.555944] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0b2d395f-d1d0-4b84-839d-704cbd0407b8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.563597] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52ca3126-e77a-d1ce-500d-dc3cb001d3e3, 'name': SearchDatastore_Task, 'duration_secs': 0.012815} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.565346] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 837.565511] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 837.566184] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5e904d44-ccc4-4b58-b734-07bd7e2bf3e6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.568364] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d6b833a3-f079-4c0e-b8a6-45b465489f57 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.574603] env[62460]: DEBUG oslo_vmware.api [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Waiting for the task: (returnval){ [ 837.574603] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]528f22d8-44be-d34f-b3b2-7c7ca37bd46e" [ 837.574603] env[62460]: _type = "Task" [ 837.574603] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.575659] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Waiting for the task: (returnval){ [ 837.575659] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52583c29-c34b-6037-9540-94dc95e8c482" [ 837.575659] env[62460]: _type = "Task" [ 837.575659] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.587135] env[62460]: DEBUG oslo_vmware.api [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]528f22d8-44be-d34f-b3b2-7c7ca37bd46e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.590769] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52583c29-c34b-6037-9540-94dc95e8c482, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.716395] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9c83d044-4a1e-4d35-a1ca-cefdd159ac66 tempest-ServerMetadataNegativeTestJSON-473879121 tempest-ServerMetadataNegativeTestJSON-473879121-project-member] Lock "db09d1f5-88cc-4dc7-9a7b-5d53d09567fa" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.049s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.806825] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313597, 'name': PowerOnVM_Task, 'duration_secs': 0.514782} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.807108] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 837.807315] env[62460]: INFO nova.compute.manager [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] Took 8.18 seconds to spawn the instance on the hypervisor. [ 837.807500] env[62460]: DEBUG nova.compute.manager [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 837.808424] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2233da9c-4bb7-4072-9e16-2092e66a0696 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.920557] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313598, 'name': CreateVM_Task, 'duration_secs': 0.333676} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.920741] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 837.921561] env[62460]: DEBUG oslo_concurrency.lockutils [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.921739] env[62460]: DEBUG oslo_concurrency.lockutils [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.922090] env[62460]: DEBUG oslo_concurrency.lockutils [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 837.922352] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-156c38dc-2729-4b5c-a7a9-647f6645b2ab {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.928534] env[62460]: DEBUG oslo_vmware.api [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 837.928534] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52f455a0-d147-791c-11d4-2f6ae34cbaef" [ 837.928534] env[62460]: _type = "Task" [ 837.928534] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.938586] env[62460]: DEBUG oslo_vmware.api [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52f455a0-d147-791c-11d4-2f6ae34cbaef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.978853] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd35464f-2a1a-4ba3-9e7f-0a7ed43a0290 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 837.979574] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b333a8f2-6015-433b-9065-9d05a64cf4fc {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.989479] env[62460]: DEBUG oslo_vmware.api [None req-cd35464f-2a1a-4ba3-9e7f-0a7ed43a0290 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Waiting for the task: (returnval){ [ 837.989479] env[62460]: value = "task-1313599" [ 837.989479] env[62460]: _type = "Task" [ 837.989479] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.997990] env[62460]: DEBUG oslo_vmware.api [None req-cd35464f-2a1a-4ba3-9e7f-0a7ed43a0290 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313599, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.042808] env[62460]: DEBUG oslo_concurrency.lockutils [req-d835444e-5f10-494a-aa54-abcc2561dfd5 req-7e954fdc-9dc9-420b-ae32-85a5de614706 service nova] Releasing lock "refresh_cache-0269dc64-d2b1-43c5-bdf7-11d97e534819" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.043116] env[62460]: DEBUG nova.compute.manager [req-d835444e-5f10-494a-aa54-abcc2561dfd5 req-7e954fdc-9dc9-420b-ae32-85a5de614706 service nova] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Received event network-changed-3515dec2-bd6c-4f96-aff1-8dd32de6f7d6 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 838.043304] env[62460]: DEBUG nova.compute.manager [req-d835444e-5f10-494a-aa54-abcc2561dfd5 req-7e954fdc-9dc9-420b-ae32-85a5de614706 service nova] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Refreshing instance network info cache due to event network-changed-3515dec2-bd6c-4f96-aff1-8dd32de6f7d6. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 838.043683] env[62460]: DEBUG oslo_concurrency.lockutils [req-d835444e-5f10-494a-aa54-abcc2561dfd5 req-7e954fdc-9dc9-420b-ae32-85a5de614706 service nova] Acquiring lock "refresh_cache-5b6e8205-003e-49c0-a73d-be2e032a8272" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.043953] env[62460]: DEBUG oslo_concurrency.lockutils [req-d835444e-5f10-494a-aa54-abcc2561dfd5 req-7e954fdc-9dc9-420b-ae32-85a5de614706 service nova] Acquired lock "refresh_cache-5b6e8205-003e-49c0-a73d-be2e032a8272" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.044228] env[62460]: DEBUG nova.network.neutron [req-d835444e-5f10-494a-aa54-abcc2561dfd5 req-7e954fdc-9dc9-420b-ae32-85a5de614706 service nova] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Refreshing network info cache for port 3515dec2-bd6c-4f96-aff1-8dd32de6f7d6 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 838.089928] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52583c29-c34b-6037-9540-94dc95e8c482, 'name': SearchDatastore_Task, 'duration_secs': 0.014629} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.094183] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.094468] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 0269dc64-d2b1-43c5-bdf7-11d97e534819/0269dc64-d2b1-43c5-bdf7-11d97e534819.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 838.094757] env[62460]: DEBUG oslo_vmware.api [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]528f22d8-44be-d34f-b3b2-7c7ca37bd46e, 'name': SearchDatastore_Task, 'duration_secs': 0.012874} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.095398] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-54281561-42ee-43d0-ab76-a0a3e70de3d1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.097557] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d1cff23b-1e23-4a31-93d2-0d4106f19aff {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.102639] env[62460]: DEBUG oslo_vmware.api [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Waiting for the task: (returnval){ [ 838.102639] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]5245bc87-7f2f-5c4b-4e07-a92b1e46a0c8" [ 838.102639] env[62460]: _type = "Task" [ 838.102639] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.107161] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Waiting for the task: (returnval){ [ 838.107161] env[62460]: value = "task-1313600" [ 838.107161] env[62460]: _type = "Task" [ 838.107161] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.115309] env[62460]: DEBUG oslo_vmware.api [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5245bc87-7f2f-5c4b-4e07-a92b1e46a0c8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.120698] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313600, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.328255] env[62460]: INFO nova.compute.manager [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] Took 32.09 seconds to build instance. [ 838.445566] env[62460]: DEBUG oslo_vmware.api [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52f455a0-d147-791c-11d4-2f6ae34cbaef, 'name': SearchDatastore_Task, 'duration_secs': 0.029582} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.446321] env[62460]: DEBUG oslo_concurrency.lockutils [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.446321] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 838.446533] env[62460]: DEBUG oslo_concurrency.lockutils [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.491952] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cae348a-e4c8-49a3-a7f1-256f008b758e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.509495] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52d1159a-2309-462a-99a2-80ea77512eb5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.513532] env[62460]: DEBUG oslo_vmware.api [None req-cd35464f-2a1a-4ba3-9e7f-0a7ed43a0290 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313599, 'name': PowerOnVM_Task, 'duration_secs': 0.482728} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.513877] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd35464f-2a1a-4ba3-9e7f-0a7ed43a0290 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 838.514081] env[62460]: DEBUG nova.compute.manager [None req-cd35464f-2a1a-4ba3-9e7f-0a7ed43a0290 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 838.515381] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e50164f8-c1ac-4dae-af3e-919f08727116 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.547422] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d968ba8d-47f5-4b75-8928-55152f16f644 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.564426] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b381e10b-37b7-44e5-971c-f166265e20d1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.584220] env[62460]: DEBUG nova.compute.provider_tree [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 838.621567] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313600, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.626235] env[62460]: DEBUG oslo_vmware.api [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5245bc87-7f2f-5c4b-4e07-a92b1e46a0c8, 'name': SearchDatastore_Task, 'duration_secs': 0.017959} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.626578] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.626805] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] 9d1a5830-f3c0-4d18-9338-16f7b6962c6a/9d1a5830-f3c0-4d18-9338-16f7b6962c6a.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 838.627131] env[62460]: DEBUG oslo_concurrency.lockutils [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.627436] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 838.627604] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-81444b2d-7a54-4286-b8c2-47cb0342a5a2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.636059] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-18662b94-568e-4a66-9dc1-bd2b7385f298 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.645187] env[62460]: DEBUG oslo_vmware.api [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Waiting for the task: (returnval){ [ 838.645187] env[62460]: value = "task-1313601" [ 838.645187] env[62460]: _type = "Task" [ 838.645187] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.650549] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 838.650775] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 838.652177] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-44535e78-fbc2-4ca5-9b4b-1aecbb934f33 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.658068] env[62460]: DEBUG oslo_vmware.api [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Task: {'id': task-1313601, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.661893] env[62460]: DEBUG oslo_vmware.api [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 838.661893] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52915080-d720-71a9-ebcd-0d3dff2cd04d" [ 838.661893] env[62460]: _type = "Task" [ 838.661893] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.672064] env[62460]: DEBUG oslo_vmware.api [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52915080-d720-71a9-ebcd-0d3dff2cd04d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.833216] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Lock "9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 104.895s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.037793] env[62460]: DEBUG nova.network.neutron [req-d835444e-5f10-494a-aa54-abcc2561dfd5 req-7e954fdc-9dc9-420b-ae32-85a5de614706 service nova] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Updated VIF entry in instance network info cache for port 3515dec2-bd6c-4f96-aff1-8dd32de6f7d6. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 839.038194] env[62460]: DEBUG nova.network.neutron [req-d835444e-5f10-494a-aa54-abcc2561dfd5 req-7e954fdc-9dc9-420b-ae32-85a5de614706 service nova] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Updating instance_info_cache with network_info: [{"id": "3515dec2-bd6c-4f96-aff1-8dd32de6f7d6", "address": "fa:16:3e:53:9c:d5", "network": {"id": "4cb03486-3bea-4c6a-a019-37486f271107", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-2094026403-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.174", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfc48e301fb04231b88ec28a062bda4e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c7d2575f-b92f-44ec-a863-634cb76631a2", "external-id": "nsx-vlan-transportzone-794", "segmentation_id": 794, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3515dec2-bd", "ovs_interfaceid": "3515dec2-bd6c-4f96-aff1-8dd32de6f7d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.088141] env[62460]: DEBUG nova.scheduler.client.report [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 839.123724] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313600, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.622343} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.123980] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 0269dc64-d2b1-43c5-bdf7-11d97e534819/0269dc64-d2b1-43c5-bdf7-11d97e534819.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 839.124225] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 839.124818] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-afdcb916-bd05-4586-b183-b63611276e84 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.133864] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Waiting for the task: (returnval){ [ 839.133864] env[62460]: value = "task-1313602" [ 839.133864] env[62460]: _type = "Task" [ 839.133864] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.145316] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313602, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.155593] env[62460]: DEBUG oslo_vmware.api [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Task: {'id': task-1313601, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.176300] env[62460]: DEBUG oslo_vmware.api [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52915080-d720-71a9-ebcd-0d3dff2cd04d, 'name': SearchDatastore_Task, 'duration_secs': 0.039618} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.177626] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c1a4a389-0468-47e3-a2e2-83e300e3b1e6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.185948] env[62460]: DEBUG oslo_vmware.api [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 839.185948] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52d0eb4f-2fb4-38c9-8182-00fc29389ddb" [ 839.185948] env[62460]: _type = "Task" [ 839.185948] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.196164] env[62460]: DEBUG oslo_vmware.api [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52d0eb4f-2fb4-38c9-8182-00fc29389ddb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.335781] env[62460]: DEBUG nova.compute.manager [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 839.545516] env[62460]: DEBUG oslo_concurrency.lockutils [req-d835444e-5f10-494a-aa54-abcc2561dfd5 req-7e954fdc-9dc9-420b-ae32-85a5de614706 service nova] Releasing lock "refresh_cache-5b6e8205-003e-49c0-a73d-be2e032a8272" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.597254] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.406s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.597254] env[62460]: DEBUG nova.compute.manager [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 839.599644] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 19.987s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.599960] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.600358] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62460) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 839.601060] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.473s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.602939] env[62460]: INFO nova.compute.claims [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 839.607186] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c047c27-db2f-44f5-a5e3-533ff7489143 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.617667] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f93291f5-ee91-4eba-98e2-8b2888ad5afa {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.642150] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03d932fc-ac03-4e5b-9740-f56073317f2a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.655799] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313602, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082726} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.657240] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2c4bae2-e788-455c-aa8e-926f7f51ee63 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.661865] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 839.666532] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fc569dd-ad1d-48ae-a98f-25d194b7c2a5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.669628] env[62460]: DEBUG oslo_vmware.api [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Task: {'id': task-1313601, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.577077} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.670855] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] 9d1a5830-f3c0-4d18-9338-16f7b6962c6a/9d1a5830-f3c0-4d18-9338-16f7b6962c6a.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 839.671609] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 839.672683] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d914f3cc-6ba0-4cef-a28d-aaa18054e331 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.720927] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] Reconfiguring VM instance instance-00000044 to attach disk [datastore1] 0269dc64-d2b1-43c5-bdf7-11d97e534819/0269dc64-d2b1-43c5-bdf7-11d97e534819.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 839.721600] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181332MB free_disk=134GB free_vcpus=48 pci_devices=None {{(pid=62460) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 839.721750] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.723041] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f1d24e2c-d552-4a82-8887-391127986d06 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.744371] env[62460]: DEBUG oslo_vmware.api [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Waiting for the task: (returnval){ [ 839.744371] env[62460]: value = "task-1313603" [ 839.744371] env[62460]: _type = "Task" [ 839.744371] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.754018] env[62460]: DEBUG oslo_vmware.api [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52d0eb4f-2fb4-38c9-8182-00fc29389ddb, 'name': SearchDatastore_Task, 'duration_secs': 0.019982} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.754018] env[62460]: DEBUG oslo_concurrency.lockutils [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.754319] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] b76f37a0-91d0-4a01-9d95-9c6586081175/b76f37a0-91d0-4a01-9d95-9c6586081175.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 839.754670] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Waiting for the task: (returnval){ [ 839.754670] env[62460]: value = "task-1313604" [ 839.754670] env[62460]: _type = "Task" [ 839.754670] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.754926] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9cb9ac0c-474e-4937-8e1a-c8229e616601 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.764610] env[62460]: DEBUG oslo_vmware.api [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Task: {'id': task-1313603, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.772941] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313604, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.777202] env[62460]: DEBUG oslo_vmware.api [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 839.777202] env[62460]: value = "task-1313605" [ 839.777202] env[62460]: _type = "Task" [ 839.777202] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.786419] env[62460]: DEBUG oslo_vmware.api [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313605, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.865679] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.107496] env[62460]: DEBUG nova.compute.utils [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 840.109159] env[62460]: DEBUG nova.compute.manager [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 840.109431] env[62460]: DEBUG nova.network.neutron [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 840.259459] env[62460]: DEBUG oslo_vmware.api [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Task: {'id': task-1313603, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075728} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.264493] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 840.265583] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dff3c546-59c8-4d94-8e32-1aefaa7870e8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.278638] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313604, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.302481] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Reconfiguring VM instance instance-00000045 to attach disk [datastore2] 9d1a5830-f3c0-4d18-9338-16f7b6962c6a/9d1a5830-f3c0-4d18-9338-16f7b6962c6a.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 840.303051] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a1c45ccb-cce6-4822-a743-728ecfc91e56 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.340798] env[62460]: DEBUG oslo_vmware.api [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313605, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.343090] env[62460]: DEBUG nova.policy [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5f69b3e3f50248ff8e57054294a04686', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0da056d93bdf40c39d6e82e457727ff6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 840.347347] env[62460]: DEBUG oslo_vmware.api [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Waiting for the task: (returnval){ [ 840.347347] env[62460]: value = "task-1313606" [ 840.347347] env[62460]: _type = "Task" [ 840.347347] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.363571] env[62460]: DEBUG oslo_vmware.api [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Task: {'id': task-1313606, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.461994] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7eaf8235-f240-4400-9d00-8b18849f1e54 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Acquiring lock "1f318a64-2c38-470b-8fae-4ba4543a5681" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.462411] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7eaf8235-f240-4400-9d00-8b18849f1e54 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Lock "1f318a64-2c38-470b-8fae-4ba4543a5681" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.462664] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7eaf8235-f240-4400-9d00-8b18849f1e54 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Acquiring lock "1f318a64-2c38-470b-8fae-4ba4543a5681-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.462943] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7eaf8235-f240-4400-9d00-8b18849f1e54 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Lock "1f318a64-2c38-470b-8fae-4ba4543a5681-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.463127] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7eaf8235-f240-4400-9d00-8b18849f1e54 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Lock "1f318a64-2c38-470b-8fae-4ba4543a5681-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.465634] env[62460]: INFO nova.compute.manager [None req-7eaf8235-f240-4400-9d00-8b18849f1e54 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Terminating instance [ 840.469770] env[62460]: DEBUG nova.compute.manager [None req-7eaf8235-f240-4400-9d00-8b18849f1e54 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 840.469770] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7eaf8235-f240-4400-9d00-8b18849f1e54 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 840.469770] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f232ed6-63b7-4233-8190-3ce19275b75a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.480577] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-7eaf8235-f240-4400-9d00-8b18849f1e54 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 840.481809] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-842ceac2-85d5-4ae8-826c-9136b3ec2ae7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.489744] env[62460]: DEBUG oslo_vmware.api [None req-7eaf8235-f240-4400-9d00-8b18849f1e54 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Waiting for the task: (returnval){ [ 840.489744] env[62460]: value = "task-1313607" [ 840.489744] env[62460]: _type = "Task" [ 840.489744] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.499579] env[62460]: DEBUG oslo_vmware.api [None req-7eaf8235-f240-4400-9d00-8b18849f1e54 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Task: {'id': task-1313607, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.612641] env[62460]: DEBUG nova.compute.manager [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 840.769392] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313604, 'name': ReconfigVM_Task, 'duration_secs': 0.577908} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.772428] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] Reconfigured VM instance instance-00000044 to attach disk [datastore1] 0269dc64-d2b1-43c5-bdf7-11d97e534819/0269dc64-d2b1-43c5-bdf7-11d97e534819.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 840.773439] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4499d393-d7e3-483f-b331-0a400efa4ffd {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.784588] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Waiting for the task: (returnval){ [ 840.784588] env[62460]: value = "task-1313608" [ 840.784588] env[62460]: _type = "Task" [ 840.784588] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.797582] env[62460]: DEBUG oslo_vmware.api [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313605, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.832167} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.798780] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] b76f37a0-91d0-4a01-9d95-9c6586081175/b76f37a0-91d0-4a01-9d95-9c6586081175.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 840.799029] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 840.799071] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1cdfcc1d-db77-4c57-bd26-d0603f1a97fd {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.805174] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313608, 'name': Rename_Task} progress is 10%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.807170] env[62460]: DEBUG oslo_vmware.api [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 840.807170] env[62460]: value = "task-1313609" [ 840.807170] env[62460]: _type = "Task" [ 840.807170] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.818212] env[62460]: DEBUG oslo_vmware.api [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313609, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.819175] env[62460]: DEBUG nova.network.neutron [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Successfully created port: 5606455d-d407-41ff-8232-841d166aa4b1 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 840.861984] env[62460]: DEBUG oslo_vmware.api [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Task: {'id': task-1313606, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.964041] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1977d7df-cfc4-48e3-84a6-ca283415c508 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.971814] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdf84d02-f836-403f-a1c7-8737974094dc {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.005287] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7cdaaf8-8bc5-42b5-ac94-89c78a54c3d4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.015191] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f40112e8-4e6d-48c3-ba59-e1d25bcf0491 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.018861] env[62460]: DEBUG oslo_vmware.api [None req-7eaf8235-f240-4400-9d00-8b18849f1e54 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Task: {'id': task-1313607, 'name': PowerOffVM_Task, 'duration_secs': 0.440324} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.019321] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-7eaf8235-f240-4400-9d00-8b18849f1e54 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 841.019321] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7eaf8235-f240-4400-9d00-8b18849f1e54 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 841.019860] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-58543bea-85ad-47a3-865a-1be77549b382 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.029462] env[62460]: DEBUG nova.compute.provider_tree [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 841.107335] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7eaf8235-f240-4400-9d00-8b18849f1e54 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 841.107335] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7eaf8235-f240-4400-9d00-8b18849f1e54 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Deleting contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 841.107335] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-7eaf8235-f240-4400-9d00-8b18849f1e54 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Deleting the datastore file [datastore1] 1f318a64-2c38-470b-8fae-4ba4543a5681 {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 841.107643] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4e4e87b9-370f-46fa-9582-4f12933a4d8e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.113734] env[62460]: DEBUG oslo_vmware.api [None req-7eaf8235-f240-4400-9d00-8b18849f1e54 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Waiting for the task: (returnval){ [ 841.113734] env[62460]: value = "task-1313611" [ 841.113734] env[62460]: _type = "Task" [ 841.113734] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.124888] env[62460]: DEBUG oslo_vmware.api [None req-7eaf8235-f240-4400-9d00-8b18849f1e54 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Task: {'id': task-1313611, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.295254] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313608, 'name': Rename_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.319955] env[62460]: DEBUG oslo_vmware.api [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313609, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080249} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.320444] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 841.321704] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebf40303-a8ca-4b88-b97d-6f66ce844eb3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.346979] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Reconfiguring VM instance instance-00000041 to attach disk [datastore2] b76f37a0-91d0-4a01-9d95-9c6586081175/b76f37a0-91d0-4a01-9d95-9c6586081175.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 841.347714] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-383bb858-8f7e-4ab8-a747-549e9837d485 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.371767] env[62460]: DEBUG oslo_vmware.api [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Task: {'id': task-1313606, 'name': ReconfigVM_Task, 'duration_secs': 0.68968} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.372354] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Reconfigured VM instance instance-00000045 to attach disk [datastore2] 9d1a5830-f3c0-4d18-9338-16f7b6962c6a/9d1a5830-f3c0-4d18-9338-16f7b6962c6a.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 841.373041] env[62460]: DEBUG oslo_vmware.api [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 841.373041] env[62460]: value = "task-1313612" [ 841.373041] env[62460]: _type = "Task" [ 841.373041] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.373240] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-39a57c5f-0966-4f69-8b97-1489d8d64c20 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.383143] env[62460]: DEBUG oslo_vmware.api [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313612, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.384679] env[62460]: DEBUG oslo_vmware.api [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Waiting for the task: (returnval){ [ 841.384679] env[62460]: value = "task-1313613" [ 841.384679] env[62460]: _type = "Task" [ 841.384679] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.392681] env[62460]: DEBUG oslo_vmware.api [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Task: {'id': task-1313613, 'name': Rename_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.532572] env[62460]: DEBUG nova.scheduler.client.report [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 841.623459] env[62460]: DEBUG oslo_vmware.api [None req-7eaf8235-f240-4400-9d00-8b18849f1e54 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Task: {'id': task-1313611, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.626756] env[62460]: DEBUG nova.compute.manager [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 841.651789] env[62460]: DEBUG nova.virt.hardware [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 841.652209] env[62460]: DEBUG nova.virt.hardware [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 841.652427] env[62460]: DEBUG nova.virt.hardware [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 841.652740] env[62460]: DEBUG nova.virt.hardware [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 841.652914] env[62460]: DEBUG nova.virt.hardware [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 841.653083] env[62460]: DEBUG nova.virt.hardware [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 841.653300] env[62460]: DEBUG nova.virt.hardware [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 841.653549] env[62460]: DEBUG nova.virt.hardware [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 841.653814] env[62460]: DEBUG nova.virt.hardware [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 841.653995] env[62460]: DEBUG nova.virt.hardware [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 841.654197] env[62460]: DEBUG nova.virt.hardware [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 841.655237] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d86522a-3001-4405-8d1f-9d56a4f48fda {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.663243] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb7b8930-0572-40af-855f-1fc944d65cb4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.772025] env[62460]: DEBUG oslo_concurrency.lockutils [None req-14cfd38a-3e6f-431b-b68a-ba371f751726 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "interface-7e6ff902-4a04-43d5-9014-38c4ec88efc4-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.772025] env[62460]: DEBUG oslo_concurrency.lockutils [None req-14cfd38a-3e6f-431b-b68a-ba371f751726 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "interface-7e6ff902-4a04-43d5-9014-38c4ec88efc4-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.772025] env[62460]: DEBUG nova.objects.instance [None req-14cfd38a-3e6f-431b-b68a-ba371f751726 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lazy-loading 'flavor' on Instance uuid 7e6ff902-4a04-43d5-9014-38c4ec88efc4 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 841.794985] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313608, 'name': Rename_Task, 'duration_secs': 0.83934} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.795279] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 841.795535] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f04d205d-35da-43c8-adb9-e185176d1c33 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.801773] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Waiting for the task: (returnval){ [ 841.801773] env[62460]: value = "task-1313614" [ 841.801773] env[62460]: _type = "Task" [ 841.801773] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.811227] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313614, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.884938] env[62460]: DEBUG oslo_vmware.api [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313612, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.892389] env[62460]: DEBUG oslo_vmware.api [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Task: {'id': task-1313613, 'name': Rename_Task, 'duration_secs': 0.392726} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.892650] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 841.892888] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2a69a423-b1ec-4eab-a2e6-6607d174433a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.898587] env[62460]: DEBUG oslo_vmware.api [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Waiting for the task: (returnval){ [ 841.898587] env[62460]: value = "task-1313615" [ 841.898587] env[62460]: _type = "Task" [ 841.898587] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.906154] env[62460]: DEBUG oslo_vmware.api [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Task: {'id': task-1313615, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.038347] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.437s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.039108] env[62460]: DEBUG nova.compute.manager [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 842.041878] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.592s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.043280] env[62460]: INFO nova.compute.claims [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 842.124334] env[62460]: DEBUG oslo_vmware.api [None req-7eaf8235-f240-4400-9d00-8b18849f1e54 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Task: {'id': task-1313611, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.551409} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.124594] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-7eaf8235-f240-4400-9d00-8b18849f1e54 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 842.124811] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7eaf8235-f240-4400-9d00-8b18849f1e54 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Deleted contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 842.125098] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7eaf8235-f240-4400-9d00-8b18849f1e54 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 842.125283] env[62460]: INFO nova.compute.manager [None req-7eaf8235-f240-4400-9d00-8b18849f1e54 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Took 1.66 seconds to destroy the instance on the hypervisor. [ 842.125598] env[62460]: DEBUG oslo.service.loopingcall [None req-7eaf8235-f240-4400-9d00-8b18849f1e54 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 842.125818] env[62460]: DEBUG nova.compute.manager [-] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 842.125942] env[62460]: DEBUG nova.network.neutron [-] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 842.273054] env[62460]: DEBUG nova.objects.instance [None req-14cfd38a-3e6f-431b-b68a-ba371f751726 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lazy-loading 'pci_requests' on Instance uuid 7e6ff902-4a04-43d5-9014-38c4ec88efc4 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 842.312440] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313614, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.385706] env[62460]: DEBUG oslo_vmware.api [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313612, 'name': ReconfigVM_Task, 'duration_secs': 0.887249} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.386072] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Reconfigured VM instance instance-00000041 to attach disk [datastore2] b76f37a0-91d0-4a01-9d95-9c6586081175/b76f37a0-91d0-4a01-9d95-9c6586081175.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 842.386832] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-22eac150-7f70-4a2b-85a9-fd2fd07177f9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.393885] env[62460]: DEBUG oslo_vmware.api [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 842.393885] env[62460]: value = "task-1313616" [ 842.393885] env[62460]: _type = "Task" [ 842.393885] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.402766] env[62460]: DEBUG oslo_vmware.api [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313616, 'name': Rename_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.410694] env[62460]: DEBUG oslo_vmware.api [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Task: {'id': task-1313615, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.463574] env[62460]: DEBUG nova.network.neutron [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Successfully updated port: 5606455d-d407-41ff-8232-841d166aa4b1 {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 842.547897] env[62460]: DEBUG nova.compute.utils [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 842.551063] env[62460]: DEBUG nova.compute.manager [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 842.551183] env[62460]: DEBUG nova.network.neutron [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 842.593647] env[62460]: DEBUG nova.policy [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '632d7ca8cb4243f990d741a3f70f422a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '10ac54db1ef54e249a077bbb3dde4242', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 842.732053] env[62460]: DEBUG nova.compute.manager [req-51852f20-4620-48b9-8f1b-24a7667ba311 req-3cc762b8-3d98-4aec-9bcd-1bdac0107d22 service nova] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Received event network-vif-plugged-5606455d-d407-41ff-8232-841d166aa4b1 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 842.732241] env[62460]: DEBUG oslo_concurrency.lockutils [req-51852f20-4620-48b9-8f1b-24a7667ba311 req-3cc762b8-3d98-4aec-9bcd-1bdac0107d22 service nova] Acquiring lock "d7d180f7-e1a8-46c8-ba8e-ca50dac474cc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.732453] env[62460]: DEBUG oslo_concurrency.lockutils [req-51852f20-4620-48b9-8f1b-24a7667ba311 req-3cc762b8-3d98-4aec-9bcd-1bdac0107d22 service nova] Lock "d7d180f7-e1a8-46c8-ba8e-ca50dac474cc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.732658] env[62460]: DEBUG oslo_concurrency.lockutils [req-51852f20-4620-48b9-8f1b-24a7667ba311 req-3cc762b8-3d98-4aec-9bcd-1bdac0107d22 service nova] Lock "d7d180f7-e1a8-46c8-ba8e-ca50dac474cc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.732802] env[62460]: DEBUG nova.compute.manager [req-51852f20-4620-48b9-8f1b-24a7667ba311 req-3cc762b8-3d98-4aec-9bcd-1bdac0107d22 service nova] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] No waiting events found dispatching network-vif-plugged-5606455d-d407-41ff-8232-841d166aa4b1 {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 842.732974] env[62460]: WARNING nova.compute.manager [req-51852f20-4620-48b9-8f1b-24a7667ba311 req-3cc762b8-3d98-4aec-9bcd-1bdac0107d22 service nova] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Received unexpected event network-vif-plugged-5606455d-d407-41ff-8232-841d166aa4b1 for instance with vm_state building and task_state spawning. [ 842.776908] env[62460]: DEBUG nova.objects.base [None req-14cfd38a-3e6f-431b-b68a-ba371f751726 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Object Instance<7e6ff902-4a04-43d5-9014-38c4ec88efc4> lazy-loaded attributes: flavor,pci_requests {{(pid=62460) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 842.777143] env[62460]: DEBUG nova.network.neutron [None req-14cfd38a-3e6f-431b-b68a-ba371f751726 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 842.816019] env[62460]: DEBUG oslo_vmware.api [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313614, 'name': PowerOnVM_Task, 'duration_secs': 0.824387} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.816019] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 842.816019] env[62460]: INFO nova.compute.manager [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] Took 8.90 seconds to spawn the instance on the hypervisor. [ 842.816019] env[62460]: DEBUG nova.compute.manager [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 842.816019] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b67d813-fe49-459c-82df-af03d1422f09 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.837830] env[62460]: DEBUG nova.network.neutron [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] Successfully created port: b1e3ec03-310d-4b60-8ead-a10ec31a5584 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 842.872131] env[62460]: DEBUG oslo_concurrency.lockutils [None req-14cfd38a-3e6f-431b-b68a-ba371f751726 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "interface-7e6ff902-4a04-43d5-9014-38c4ec88efc4-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.103s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.906708] env[62460]: DEBUG oslo_vmware.api [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313616, 'name': Rename_Task, 'duration_secs': 0.177455} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.910904] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 842.910904] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-acda334e-c052-4328-8f64-bc002f43c87f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.917031] env[62460]: DEBUG oslo_vmware.api [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Task: {'id': task-1313615, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.918247] env[62460]: DEBUG oslo_vmware.api [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 842.918247] env[62460]: value = "task-1313617" [ 842.918247] env[62460]: _type = "Task" [ 842.918247] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.925583] env[62460]: DEBUG oslo_vmware.api [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313617, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.970384] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "refresh_cache-d7d180f7-e1a8-46c8-ba8e-ca50dac474cc" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.970497] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquired lock "refresh_cache-d7d180f7-e1a8-46c8-ba8e-ca50dac474cc" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.970628] env[62460]: DEBUG nova.network.neutron [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 843.049798] env[62460]: DEBUG nova.compute.manager [req-16b48f55-663a-46ac-a0de-e524c476f756 req-817f77fb-47ea-4715-9566-4d72aeb189f9 service nova] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Received event network-vif-deleted-51efbac7-c441-4555-9a3c-d42167d75fcc {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 843.049933] env[62460]: INFO nova.compute.manager [req-16b48f55-663a-46ac-a0de-e524c476f756 req-817f77fb-47ea-4715-9566-4d72aeb189f9 service nova] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Neutron deleted interface 51efbac7-c441-4555-9a3c-d42167d75fcc; detaching it from the instance and deleting it from the info cache [ 843.050342] env[62460]: DEBUG nova.network.neutron [req-16b48f55-663a-46ac-a0de-e524c476f756 req-817f77fb-47ea-4715-9566-4d72aeb189f9 service nova] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.054718] env[62460]: DEBUG nova.compute.manager [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 843.263627] env[62460]: DEBUG nova.network.neutron [-] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.312678] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dbcdb89-800a-4cc4-9b75-c48b9b1ab1b9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.321019] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c85ea56-3706-4a4b-b881-0016af71e96a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.360318] env[62460]: INFO nova.compute.manager [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] Took 33.19 seconds to build instance. [ 843.362038] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-283d73d9-ed6f-4b8c-8247-a66adcd3253a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.372823] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d61b2b33-8a6d-49c9-a94b-f6895fa617c6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.394984] env[62460]: DEBUG nova.compute.provider_tree [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 843.417491] env[62460]: DEBUG oslo_vmware.api [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Task: {'id': task-1313615, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.428068] env[62460]: DEBUG oslo_vmware.api [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313617, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.517087] env[62460]: DEBUG nova.network.neutron [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 843.552609] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3fcf3a79-f148-4458-8bb0-78d0fb698d26 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.565019] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8de0b1e0-b1be-4207-8056-2a0213803e21 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.596552] env[62460]: DEBUG nova.compute.manager [req-16b48f55-663a-46ac-a0de-e524c476f756 req-817f77fb-47ea-4715-9566-4d72aeb189f9 service nova] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Detach interface failed, port_id=51efbac7-c441-4555-9a3c-d42167d75fcc, reason: Instance 1f318a64-2c38-470b-8fae-4ba4543a5681 could not be found. {{(pid=62460) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 843.738515] env[62460]: DEBUG nova.network.neutron [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Updating instance_info_cache with network_info: [{"id": "5606455d-d407-41ff-8232-841d166aa4b1", "address": "fa:16:3e:6d:35:05", "network": {"id": "de789887-9cf9-4614-a7e1-dab430aaabaf", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-656400207-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0da056d93bdf40c39d6e82e457727ff6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5116f690-f825-4fee-8a47-42b073e716c5", "external-id": "nsx-vlan-transportzone-692", "segmentation_id": 692, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5606455d-d4", "ovs_interfaceid": "5606455d-d407-41ff-8232-841d166aa4b1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.766681] env[62460]: INFO nova.compute.manager [-] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Took 1.64 seconds to deallocate network for instance. [ 843.862340] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6fb7f11d-fd0f-4a54-a832-4885e82a3984 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Lock "0269dc64-d2b1-43c5-bdf7-11d97e534819" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 109.897s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.900501] env[62460]: DEBUG nova.scheduler.client.report [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 843.918929] env[62460]: DEBUG oslo_vmware.api [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Task: {'id': task-1313615, 'name': PowerOnVM_Task, 'duration_secs': 1.677606} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.919805] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 843.920029] env[62460]: INFO nova.compute.manager [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Took 7.54 seconds to spawn the instance on the hypervisor. [ 843.920225] env[62460]: DEBUG nova.compute.manager [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 843.921059] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3630797b-00de-4b59-ad93-1c1952700dad {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.932068] env[62460]: DEBUG oslo_vmware.api [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313617, 'name': PowerOnVM_Task, 'duration_secs': 0.846699} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.933563] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 843.933794] env[62460]: DEBUG nova.compute.manager [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 843.937935] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d6f53fc-8441-4519-8e10-431f7367c2a9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.081542] env[62460]: DEBUG nova.compute.manager [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 844.108052] env[62460]: DEBUG nova.virt.hardware [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 844.108314] env[62460]: DEBUG nova.virt.hardware [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 844.108474] env[62460]: DEBUG nova.virt.hardware [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 844.108655] env[62460]: DEBUG nova.virt.hardware [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 844.108801] env[62460]: DEBUG nova.virt.hardware [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 844.108950] env[62460]: DEBUG nova.virt.hardware [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 844.109179] env[62460]: DEBUG nova.virt.hardware [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 844.109361] env[62460]: DEBUG nova.virt.hardware [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 844.109504] env[62460]: DEBUG nova.virt.hardware [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 844.109666] env[62460]: DEBUG nova.virt.hardware [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 844.109836] env[62460]: DEBUG nova.virt.hardware [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 844.110972] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-884facd7-a92d-47b4-afcc-d64499aa3407 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.119609] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6580c6b-56ef-4567-a3f4-5d288baf713a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.241797] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Releasing lock "refresh_cache-d7d180f7-e1a8-46c8-ba8e-ca50dac474cc" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.242181] env[62460]: DEBUG nova.compute.manager [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Instance network_info: |[{"id": "5606455d-d407-41ff-8232-841d166aa4b1", "address": "fa:16:3e:6d:35:05", "network": {"id": "de789887-9cf9-4614-a7e1-dab430aaabaf", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-656400207-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0da056d93bdf40c39d6e82e457727ff6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5116f690-f825-4fee-8a47-42b073e716c5", "external-id": "nsx-vlan-transportzone-692", "segmentation_id": 692, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5606455d-d4", "ovs_interfaceid": "5606455d-d407-41ff-8232-841d166aa4b1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 844.242606] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6d:35:05', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5116f690-f825-4fee-8a47-42b073e716c5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5606455d-d407-41ff-8232-841d166aa4b1', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 844.250226] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Creating folder: Project (0da056d93bdf40c39d6e82e457727ff6). Parent ref: group-v281134. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 844.250557] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7732b8de-7a69-4f81-9411-6b18b9caf5ff {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.261903] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Created folder: Project (0da056d93bdf40c39d6e82e457727ff6) in parent group-v281134. [ 844.262125] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Creating folder: Instances. Parent ref: group-v281194. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 844.262376] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3e4276d6-827b-427f-890a-ba899e9d690c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.272088] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Created folder: Instances in parent group-v281194. [ 844.272366] env[62460]: DEBUG oslo.service.loopingcall [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 844.272522] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 844.272742] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3119fad9-5951-40ea-9445-b137b248e9e1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.290510] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7eaf8235-f240-4400-9d00-8b18849f1e54 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.292376] env[62460]: DEBUG oslo_concurrency.lockutils [None req-810313b0-6843-4505-a7de-48602dfc2f7b tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Acquiring lock "fde12685-d3b9-46a0-8931-25b904d4f21e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.292591] env[62460]: DEBUG oslo_concurrency.lockutils [None req-810313b0-6843-4505-a7de-48602dfc2f7b tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Lock "fde12685-d3b9-46a0-8931-25b904d4f21e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.292794] env[62460]: DEBUG oslo_concurrency.lockutils [None req-810313b0-6843-4505-a7de-48602dfc2f7b tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Acquiring lock "fde12685-d3b9-46a0-8931-25b904d4f21e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.293096] env[62460]: DEBUG oslo_concurrency.lockutils [None req-810313b0-6843-4505-a7de-48602dfc2f7b tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Lock "fde12685-d3b9-46a0-8931-25b904d4f21e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.293154] env[62460]: DEBUG oslo_concurrency.lockutils [None req-810313b0-6843-4505-a7de-48602dfc2f7b tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Lock "fde12685-d3b9-46a0-8931-25b904d4f21e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.296177] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 844.296177] env[62460]: value = "task-1313620" [ 844.296177] env[62460]: _type = "Task" [ 844.296177] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.296646] env[62460]: INFO nova.compute.manager [None req-810313b0-6843-4505-a7de-48602dfc2f7b tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] Terminating instance [ 844.301817] env[62460]: DEBUG nova.compute.manager [None req-810313b0-6843-4505-a7de-48602dfc2f7b tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 844.301817] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-810313b0-6843-4505-a7de-48602dfc2f7b tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 844.302551] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f626077-fca9-427d-8b40-4af42910831c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.317802] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313620, 'name': CreateVM_Task} progress is 6%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.318112] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-810313b0-6843-4505-a7de-48602dfc2f7b tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 844.318346] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-950d1d5d-4c98-4ec3-a299-866c5317a410 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.325358] env[62460]: DEBUG oslo_vmware.api [None req-810313b0-6843-4505-a7de-48602dfc2f7b tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Waiting for the task: (returnval){ [ 844.325358] env[62460]: value = "task-1313621" [ 844.325358] env[62460]: _type = "Task" [ 844.325358] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.333043] env[62460]: DEBUG oslo_vmware.api [None req-810313b0-6843-4505-a7de-48602dfc2f7b tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313621, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.405788] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.364s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.406348] env[62460]: DEBUG nova.compute.manager [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 844.409634] env[62460]: DEBUG oslo_concurrency.lockutils [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.360s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.411014] env[62460]: INFO nova.compute.claims [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 844.458633] env[62460]: DEBUG nova.network.neutron [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] Successfully updated port: b1e3ec03-310d-4b60-8ead-a10ec31a5584 {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 844.462551] env[62460]: INFO nova.compute.manager [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Took 32.09 seconds to build instance. [ 844.469439] env[62460]: DEBUG oslo_concurrency.lockutils [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.536060] env[62460]: DEBUG oslo_concurrency.lockutils [None req-af702ee7-e128-4206-aea5-9106143dcc05 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Acquiring lock "9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.536339] env[62460]: DEBUG oslo_concurrency.lockutils [None req-af702ee7-e128-4206-aea5-9106143dcc05 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Lock "9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.536557] env[62460]: DEBUG oslo_concurrency.lockutils [None req-af702ee7-e128-4206-aea5-9106143dcc05 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Acquiring lock "9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.536783] env[62460]: DEBUG oslo_concurrency.lockutils [None req-af702ee7-e128-4206-aea5-9106143dcc05 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Lock "9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.536961] env[62460]: DEBUG oslo_concurrency.lockutils [None req-af702ee7-e128-4206-aea5-9106143dcc05 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Lock "9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.539042] env[62460]: INFO nova.compute.manager [None req-af702ee7-e128-4206-aea5-9106143dcc05 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] Terminating instance [ 844.541181] env[62460]: DEBUG nova.compute.manager [None req-af702ee7-e128-4206-aea5-9106143dcc05 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 844.541181] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-af702ee7-e128-4206-aea5-9106143dcc05 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 844.541941] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-570ee8f0-4af8-4721-b13c-51f7147c314e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.550471] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-af702ee7-e128-4206-aea5-9106143dcc05 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 844.551301] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0870a558-94dd-4d62-865a-fd6a571bee0a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.556489] env[62460]: DEBUG oslo_vmware.api [None req-af702ee7-e128-4206-aea5-9106143dcc05 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Waiting for the task: (returnval){ [ 844.556489] env[62460]: value = "task-1313622" [ 844.556489] env[62460]: _type = "Task" [ 844.556489] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.564665] env[62460]: DEBUG oslo_vmware.api [None req-af702ee7-e128-4206-aea5-9106143dcc05 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313622, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.807981] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313620, 'name': CreateVM_Task, 'duration_secs': 0.453117} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.808220] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 844.809017] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.809262] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.809635] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 844.809923] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1c548a78-937a-4582-95a1-704c97837a53 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.814617] env[62460]: DEBUG oslo_vmware.api [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 844.814617] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52ce3c03-4a2d-bf51-2dbd-b434bb975d8b" [ 844.814617] env[62460]: _type = "Task" [ 844.814617] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.817507] env[62460]: DEBUG nova.compute.manager [req-74938e89-a9d0-47b6-9b1a-86e2825bc273 req-1bd8eaeb-75dd-4bdf-ae42-1a30ee118627 service nova] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Received event network-changed-5606455d-d407-41ff-8232-841d166aa4b1 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 844.817731] env[62460]: DEBUG nova.compute.manager [req-74938e89-a9d0-47b6-9b1a-86e2825bc273 req-1bd8eaeb-75dd-4bdf-ae42-1a30ee118627 service nova] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Refreshing instance network info cache due to event network-changed-5606455d-d407-41ff-8232-841d166aa4b1. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 844.817993] env[62460]: DEBUG oslo_concurrency.lockutils [req-74938e89-a9d0-47b6-9b1a-86e2825bc273 req-1bd8eaeb-75dd-4bdf-ae42-1a30ee118627 service nova] Acquiring lock "refresh_cache-d7d180f7-e1a8-46c8-ba8e-ca50dac474cc" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.818215] env[62460]: DEBUG oslo_concurrency.lockutils [req-74938e89-a9d0-47b6-9b1a-86e2825bc273 req-1bd8eaeb-75dd-4bdf-ae42-1a30ee118627 service nova] Acquired lock "refresh_cache-d7d180f7-e1a8-46c8-ba8e-ca50dac474cc" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.818402] env[62460]: DEBUG nova.network.neutron [req-74938e89-a9d0-47b6-9b1a-86e2825bc273 req-1bd8eaeb-75dd-4bdf-ae42-1a30ee118627 service nova] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Refreshing network info cache for port 5606455d-d407-41ff-8232-841d166aa4b1 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 844.830599] env[62460]: DEBUG oslo_vmware.api [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52ce3c03-4a2d-bf51-2dbd-b434bb975d8b, 'name': SearchDatastore_Task, 'duration_secs': 0.008735} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.834713] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.835071] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 844.835350] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.835505] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.835746] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 844.836401] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4c7dc525-8bbf-41a9-8214-23269c1b88bf {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.842907] env[62460]: DEBUG oslo_concurrency.lockutils [None req-53ff2f89-d24e-47c8-a414-09b17968e3cd tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Acquiring lock "0269dc64-d2b1-43c5-bdf7-11d97e534819" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.843153] env[62460]: DEBUG oslo_concurrency.lockutils [None req-53ff2f89-d24e-47c8-a414-09b17968e3cd tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Lock "0269dc64-d2b1-43c5-bdf7-11d97e534819" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.843397] env[62460]: DEBUG oslo_concurrency.lockutils [None req-53ff2f89-d24e-47c8-a414-09b17968e3cd tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Acquiring lock "0269dc64-d2b1-43c5-bdf7-11d97e534819-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.843585] env[62460]: DEBUG oslo_concurrency.lockutils [None req-53ff2f89-d24e-47c8-a414-09b17968e3cd tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Lock "0269dc64-d2b1-43c5-bdf7-11d97e534819-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.843755] env[62460]: DEBUG oslo_concurrency.lockutils [None req-53ff2f89-d24e-47c8-a414-09b17968e3cd tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Lock "0269dc64-d2b1-43c5-bdf7-11d97e534819-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.845443] env[62460]: DEBUG oslo_vmware.api [None req-810313b0-6843-4505-a7de-48602dfc2f7b tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313621, 'name': PowerOffVM_Task, 'duration_secs': 0.231356} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.846729] env[62460]: INFO nova.compute.manager [None req-53ff2f89-d24e-47c8-a414-09b17968e3cd tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] Terminating instance [ 844.848249] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-810313b0-6843-4505-a7de-48602dfc2f7b tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 844.848425] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-810313b0-6843-4505-a7de-48602dfc2f7b tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 844.848680] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 844.848840] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 844.849840] env[62460]: DEBUG nova.compute.manager [None req-53ff2f89-d24e-47c8-a414-09b17968e3cd tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 844.850039] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-53ff2f89-d24e-47c8-a414-09b17968e3cd tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 844.850262] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b531816c-2965-41e1-8c6f-67f4c3285b24 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.851662] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fd74430a-1e2e-4c8e-bdd8-6a8af608be5e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.854114] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02a22747-be3d-445e-a7dc-3c1d8214de97 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.860091] env[62460]: DEBUG oslo_vmware.api [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 844.860091] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52ef2a29-9fe0-3cde-a67b-a604e0fe7a9d" [ 844.860091] env[62460]: _type = "Task" [ 844.860091] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.862038] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-53ff2f89-d24e-47c8-a414-09b17968e3cd tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 844.864616] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ea616102-6708-4651-99ec-613de674ba86 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.871044] env[62460]: DEBUG oslo_vmware.api [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52ef2a29-9fe0-3cde-a67b-a604e0fe7a9d, 'name': SearchDatastore_Task, 'duration_secs': 0.007589} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.872776] env[62460]: DEBUG oslo_vmware.api [None req-53ff2f89-d24e-47c8-a414-09b17968e3cd tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Waiting for the task: (returnval){ [ 844.872776] env[62460]: value = "task-1313624" [ 844.872776] env[62460]: _type = "Task" [ 844.872776] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.873010] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1af89344-e219-47f8-bfb6-de5999870203 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.884315] env[62460]: DEBUG oslo_vmware.api [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 844.884315] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52eb1e2c-c931-179f-c070-f541ede505df" [ 844.884315] env[62460]: _type = "Task" [ 844.884315] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.887357] env[62460]: DEBUG oslo_vmware.api [None req-53ff2f89-d24e-47c8-a414-09b17968e3cd tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313624, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.895127] env[62460]: DEBUG oslo_vmware.api [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52eb1e2c-c931-179f-c070-f541ede505df, 'name': SearchDatastore_Task, 'duration_secs': 0.008908} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.895374] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.895643] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] d7d180f7-e1a8-46c8-ba8e-ca50dac474cc/d7d180f7-e1a8-46c8-ba8e-ca50dac474cc.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 844.895895] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d37119f3-96df-4e8e-977d-ce96bf6140d4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.901990] env[62460]: DEBUG oslo_vmware.api [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 844.901990] env[62460]: value = "task-1313625" [ 844.901990] env[62460]: _type = "Task" [ 844.901990] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.909344] env[62460]: DEBUG oslo_vmware.api [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313625, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.915776] env[62460]: DEBUG nova.compute.utils [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 844.919466] env[62460]: DEBUG nova.compute.manager [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 844.919466] env[62460]: DEBUG nova.network.neutron [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 844.922486] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-810313b0-6843-4505-a7de-48602dfc2f7b tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 844.922690] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-810313b0-6843-4505-a7de-48602dfc2f7b tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] Deleting contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 844.922871] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-810313b0-6843-4505-a7de-48602dfc2f7b tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Deleting the datastore file [datastore2] fde12685-d3b9-46a0-8931-25b904d4f21e {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 844.923435] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-92b89c7c-2f73-43b2-9a0f-c958c22ead8b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.930705] env[62460]: DEBUG oslo_vmware.api [None req-810313b0-6843-4505-a7de-48602dfc2f7b tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Waiting for the task: (returnval){ [ 844.930705] env[62460]: value = "task-1313626" [ 844.930705] env[62460]: _type = "Task" [ 844.930705] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.940301] env[62460]: DEBUG oslo_vmware.api [None req-810313b0-6843-4505-a7de-48602dfc2f7b tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313626, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.965219] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "refresh_cache-16d40829-ba6e-4193-98b5-fff94c066bcb" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.965384] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquired lock "refresh_cache-16d40829-ba6e-4193-98b5-fff94c066bcb" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.965541] env[62460]: DEBUG nova.network.neutron [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 844.968877] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1af5b63a-aa25-4910-b759-1e80020a39ed tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Lock "9d1a5830-f3c0-4d18-9338-16f7b6962c6a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 110.388s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.994134] env[62460]: DEBUG nova.policy [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4d95c2a24f0841de8990410e24f325e9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '11491a5610734d96bac82035c46bd5d1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 845.008031] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a781a4a5-2201-450d-aedf-b182703d796e tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "interface-7e6ff902-4a04-43d5-9014-38c4ec88efc4-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.008389] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a781a4a5-2201-450d-aedf-b182703d796e tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "interface-7e6ff902-4a04-43d5-9014-38c4ec88efc4-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.008766] env[62460]: DEBUG nova.objects.instance [None req-a781a4a5-2201-450d-aedf-b182703d796e tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lazy-loading 'flavor' on Instance uuid 7e6ff902-4a04-43d5-9014-38c4ec88efc4 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 845.067441] env[62460]: DEBUG oslo_vmware.api [None req-af702ee7-e128-4206-aea5-9106143dcc05 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313622, 'name': PowerOffVM_Task, 'duration_secs': 0.290159} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.067859] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-af702ee7-e128-4206-aea5-9106143dcc05 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 845.068149] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-af702ee7-e128-4206-aea5-9106143dcc05 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 845.068502] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5c6bda60-9117-4ed4-a195-73d3889ba672 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.301540] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-af702ee7-e128-4206-aea5-9106143dcc05 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 845.301788] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-af702ee7-e128-4206-aea5-9106143dcc05 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] Deleting contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 845.301975] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-af702ee7-e128-4206-aea5-9106143dcc05 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Deleting the datastore file [datastore1] 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 845.302272] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-aeb98332-bd7a-4505-b331-01b9626f13e5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.310071] env[62460]: DEBUG oslo_vmware.api [None req-af702ee7-e128-4206-aea5-9106143dcc05 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Waiting for the task: (returnval){ [ 845.310071] env[62460]: value = "task-1313628" [ 845.310071] env[62460]: _type = "Task" [ 845.310071] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.320447] env[62460]: DEBUG oslo_vmware.api [None req-af702ee7-e128-4206-aea5-9106143dcc05 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313628, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.365874] env[62460]: DEBUG nova.network.neutron [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Successfully created port: c6936c73-cdb0-4475-b00b-de47fc29e7cd {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 845.387152] env[62460]: DEBUG oslo_vmware.api [None req-53ff2f89-d24e-47c8-a414-09b17968e3cd tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313624, 'name': PowerOffVM_Task, 'duration_secs': 0.174501} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.387439] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-53ff2f89-d24e-47c8-a414-09b17968e3cd tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 845.387610] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-53ff2f89-d24e-47c8-a414-09b17968e3cd tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 845.387884] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3682d69b-dde3-4bd7-980d-79c03eb46b1b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.416020] env[62460]: DEBUG oslo_vmware.api [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313625, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.495796} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.416516] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] d7d180f7-e1a8-46c8-ba8e-ca50dac474cc/d7d180f7-e1a8-46c8-ba8e-ca50dac474cc.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 845.416744] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 845.417033] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-41eb8b70-4bb9-4b3a-adff-4454efbd2531 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.419511] env[62460]: DEBUG nova.compute.manager [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 845.430423] env[62460]: DEBUG oslo_vmware.api [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 845.430423] env[62460]: value = "task-1313630" [ 845.430423] env[62460]: _type = "Task" [ 845.430423] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.454877] env[62460]: DEBUG oslo_vmware.api [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313630, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.459544] env[62460]: DEBUG oslo_vmware.api [None req-810313b0-6843-4505-a7de-48602dfc2f7b tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313626, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.210687} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.459687] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-810313b0-6843-4505-a7de-48602dfc2f7b tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 845.463020] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-810313b0-6843-4505-a7de-48602dfc2f7b tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] Deleted contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 845.463020] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-810313b0-6843-4505-a7de-48602dfc2f7b tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 845.463020] env[62460]: INFO nova.compute.manager [None req-810313b0-6843-4505-a7de-48602dfc2f7b tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] Took 1.16 seconds to destroy the instance on the hypervisor. [ 845.463020] env[62460]: DEBUG oslo.service.loopingcall [None req-810313b0-6843-4505-a7de-48602dfc2f7b tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 845.463247] env[62460]: DEBUG nova.compute.manager [-] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 845.463326] env[62460]: DEBUG nova.network.neutron [-] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 845.466320] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-53ff2f89-d24e-47c8-a414-09b17968e3cd tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 845.466517] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-53ff2f89-d24e-47c8-a414-09b17968e3cd tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] Deleting contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 845.466702] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-53ff2f89-d24e-47c8-a414-09b17968e3cd tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Deleting the datastore file [datastore1] 0269dc64-d2b1-43c5-bdf7-11d97e534819 {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 845.467845] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c1c1f49d-9a88-4070-be88-24ab06b451b5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.475479] env[62460]: DEBUG oslo_vmware.api [None req-53ff2f89-d24e-47c8-a414-09b17968e3cd tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Waiting for the task: (returnval){ [ 845.475479] env[62460]: value = "task-1313631" [ 845.475479] env[62460]: _type = "Task" [ 845.475479] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.488587] env[62460]: DEBUG oslo_vmware.api [None req-53ff2f89-d24e-47c8-a414-09b17968e3cd tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313631, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.547372] env[62460]: DEBUG nova.network.neutron [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 845.667238] env[62460]: DEBUG oslo_concurrency.lockutils [None req-759824f9-86b1-41cc-87e9-b267e31f479b tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "b76f37a0-91d0-4a01-9d95-9c6586081175" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.667494] env[62460]: DEBUG oslo_concurrency.lockutils [None req-759824f9-86b1-41cc-87e9-b267e31f479b tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "b76f37a0-91d0-4a01-9d95-9c6586081175" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.667703] env[62460]: DEBUG oslo_concurrency.lockutils [None req-759824f9-86b1-41cc-87e9-b267e31f479b tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "b76f37a0-91d0-4a01-9d95-9c6586081175-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.667893] env[62460]: DEBUG oslo_concurrency.lockutils [None req-759824f9-86b1-41cc-87e9-b267e31f479b tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "b76f37a0-91d0-4a01-9d95-9c6586081175-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.668074] env[62460]: DEBUG oslo_concurrency.lockutils [None req-759824f9-86b1-41cc-87e9-b267e31f479b tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "b76f37a0-91d0-4a01-9d95-9c6586081175-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.670199] env[62460]: INFO nova.compute.manager [None req-759824f9-86b1-41cc-87e9-b267e31f479b tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Terminating instance [ 845.674667] env[62460]: DEBUG nova.compute.manager [None req-759824f9-86b1-41cc-87e9-b267e31f479b tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 845.674869] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-759824f9-86b1-41cc-87e9-b267e31f479b tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 845.678802] env[62460]: DEBUG nova.objects.instance [None req-a781a4a5-2201-450d-aedf-b182703d796e tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lazy-loading 'pci_requests' on Instance uuid 7e6ff902-4a04-43d5-9014-38c4ec88efc4 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 845.680450] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3edd2a8c-f27d-43dd-a0d7-fe97a9d9692c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.690677] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-759824f9-86b1-41cc-87e9-b267e31f479b tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 845.693382] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-969dd52a-179c-4042-b70d-428a73b44e0e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.702978] env[62460]: DEBUG oslo_vmware.api [None req-759824f9-86b1-41cc-87e9-b267e31f479b tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 845.702978] env[62460]: value = "task-1313632" [ 845.702978] env[62460]: _type = "Task" [ 845.702978] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.715121] env[62460]: DEBUG oslo_vmware.api [None req-759824f9-86b1-41cc-87e9-b267e31f479b tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313632, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.726967] env[62460]: DEBUG nova.network.neutron [req-74938e89-a9d0-47b6-9b1a-86e2825bc273 req-1bd8eaeb-75dd-4bdf-ae42-1a30ee118627 service nova] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Updated VIF entry in instance network info cache for port 5606455d-d407-41ff-8232-841d166aa4b1. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 845.727367] env[62460]: DEBUG nova.network.neutron [req-74938e89-a9d0-47b6-9b1a-86e2825bc273 req-1bd8eaeb-75dd-4bdf-ae42-1a30ee118627 service nova] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Updating instance_info_cache with network_info: [{"id": "5606455d-d407-41ff-8232-841d166aa4b1", "address": "fa:16:3e:6d:35:05", "network": {"id": "de789887-9cf9-4614-a7e1-dab430aaabaf", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-656400207-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0da056d93bdf40c39d6e82e457727ff6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5116f690-f825-4fee-8a47-42b073e716c5", "external-id": "nsx-vlan-transportzone-692", "segmentation_id": 692, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5606455d-d4", "ovs_interfaceid": "5606455d-d407-41ff-8232-841d166aa4b1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.733459] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dedc67a-7f3c-491e-b3fe-6c9b3e00ba22 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.741688] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67daccf8-e8fb-451f-ab5b-094f42f0f2e7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.776262] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4b6bb67-a23e-4c92-bdd9-4ddeca4f83ea {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.785235] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dfc5e2b-f9e4-4296-b824-66f41ec88352 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.791129] env[62460]: DEBUG nova.network.neutron [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] Updating instance_info_cache with network_info: [{"id": "b1e3ec03-310d-4b60-8ead-a10ec31a5584", "address": "fa:16:3e:7c:f9:2b", "network": {"id": "c8b3d824-c8f0-4f7b-a91f-1987e04a5f09", "bridge": "br-int", "label": "tempest-ServersTestJSON-1769082812-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10ac54db1ef54e249a077bbb3dde4242", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "255460d5-71d4-4bfd-87f1-acc10085db7f", "external-id": "nsx-vlan-transportzone-152", "segmentation_id": 152, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb1e3ec03-31", "ovs_interfaceid": "b1e3ec03-310d-4b60-8ead-a10ec31a5584", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.803594] env[62460]: DEBUG nova.compute.provider_tree [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 845.821424] env[62460]: DEBUG oslo_vmware.api [None req-af702ee7-e128-4206-aea5-9106143dcc05 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313628, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.942293] env[62460]: DEBUG oslo_vmware.api [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313630, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080229} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.942293] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 845.943893] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5dc37e9-15c3-414e-a28e-64e68ec98620 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.966880] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Reconfiguring VM instance instance-00000046 to attach disk [datastore2] d7d180f7-e1a8-46c8-ba8e-ca50dac474cc/d7d180f7-e1a8-46c8-ba8e-ca50dac474cc.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 845.969034] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9792e7e0-f838-4367-b3f4-293a6b626d21 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.991592] env[62460]: DEBUG oslo_vmware.api [None req-53ff2f89-d24e-47c8-a414-09b17968e3cd tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313631, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.994949] env[62460]: DEBUG oslo_vmware.api [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 845.994949] env[62460]: value = "task-1313633" [ 845.994949] env[62460]: _type = "Task" [ 845.994949] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.002874] env[62460]: DEBUG oslo_vmware.api [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313633, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.185600] env[62460]: DEBUG nova.objects.base [None req-a781a4a5-2201-450d-aedf-b182703d796e tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Object Instance<7e6ff902-4a04-43d5-9014-38c4ec88efc4> lazy-loaded attributes: flavor,pci_requests {{(pid=62460) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 846.185861] env[62460]: DEBUG nova.network.neutron [None req-a781a4a5-2201-450d-aedf-b182703d796e tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 846.212961] env[62460]: DEBUG oslo_vmware.api [None req-759824f9-86b1-41cc-87e9-b267e31f479b tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313632, 'name': PowerOffVM_Task, 'duration_secs': 0.438599} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.213252] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-759824f9-86b1-41cc-87e9-b267e31f479b tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 846.213252] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-759824f9-86b1-41cc-87e9-b267e31f479b tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 846.213834] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f1abade4-41bc-43ad-bf4f-0654d282d31d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.230946] env[62460]: DEBUG oslo_concurrency.lockutils [req-74938e89-a9d0-47b6-9b1a-86e2825bc273 req-1bd8eaeb-75dd-4bdf-ae42-1a30ee118627 service nova] Releasing lock "refresh_cache-d7d180f7-e1a8-46c8-ba8e-ca50dac474cc" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.231237] env[62460]: DEBUG nova.compute.manager [req-74938e89-a9d0-47b6-9b1a-86e2825bc273 req-1bd8eaeb-75dd-4bdf-ae42-1a30ee118627 service nova] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] Received event network-vif-plugged-b1e3ec03-310d-4b60-8ead-a10ec31a5584 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 846.231538] env[62460]: DEBUG oslo_concurrency.lockutils [req-74938e89-a9d0-47b6-9b1a-86e2825bc273 req-1bd8eaeb-75dd-4bdf-ae42-1a30ee118627 service nova] Acquiring lock "16d40829-ba6e-4193-98b5-fff94c066bcb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.231939] env[62460]: DEBUG oslo_concurrency.lockutils [req-74938e89-a9d0-47b6-9b1a-86e2825bc273 req-1bd8eaeb-75dd-4bdf-ae42-1a30ee118627 service nova] Lock "16d40829-ba6e-4193-98b5-fff94c066bcb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.232633] env[62460]: DEBUG oslo_concurrency.lockutils [req-74938e89-a9d0-47b6-9b1a-86e2825bc273 req-1bd8eaeb-75dd-4bdf-ae42-1a30ee118627 service nova] Lock "16d40829-ba6e-4193-98b5-fff94c066bcb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.232633] env[62460]: DEBUG nova.compute.manager [req-74938e89-a9d0-47b6-9b1a-86e2825bc273 req-1bd8eaeb-75dd-4bdf-ae42-1a30ee118627 service nova] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] No waiting events found dispatching network-vif-plugged-b1e3ec03-310d-4b60-8ead-a10ec31a5584 {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 846.232760] env[62460]: WARNING nova.compute.manager [req-74938e89-a9d0-47b6-9b1a-86e2825bc273 req-1bd8eaeb-75dd-4bdf-ae42-1a30ee118627 service nova] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] Received unexpected event network-vif-plugged-b1e3ec03-310d-4b60-8ead-a10ec31a5584 for instance with vm_state building and task_state spawning. [ 846.233535] env[62460]: DEBUG nova.compute.manager [req-74938e89-a9d0-47b6-9b1a-86e2825bc273 req-1bd8eaeb-75dd-4bdf-ae42-1a30ee118627 service nova] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] Received event network-changed-b1e3ec03-310d-4b60-8ead-a10ec31a5584 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 846.233535] env[62460]: DEBUG nova.compute.manager [req-74938e89-a9d0-47b6-9b1a-86e2825bc273 req-1bd8eaeb-75dd-4bdf-ae42-1a30ee118627 service nova] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] Refreshing instance network info cache due to event network-changed-b1e3ec03-310d-4b60-8ead-a10ec31a5584. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 846.233535] env[62460]: DEBUG oslo_concurrency.lockutils [req-74938e89-a9d0-47b6-9b1a-86e2825bc273 req-1bd8eaeb-75dd-4bdf-ae42-1a30ee118627 service nova] Acquiring lock "refresh_cache-16d40829-ba6e-4193-98b5-fff94c066bcb" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.242404] env[62460]: DEBUG nova.policy [None req-a781a4a5-2201-450d-aedf-b182703d796e tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd7b38fb0a53f4d078ae86d68e3eb6268', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '53fa06008e9f43488362895e7a143700', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 846.281861] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-759824f9-86b1-41cc-87e9-b267e31f479b tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 846.282072] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-759824f9-86b1-41cc-87e9-b267e31f479b tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Deleting contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 846.282493] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-759824f9-86b1-41cc-87e9-b267e31f479b tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Deleting the datastore file [datastore2] b76f37a0-91d0-4a01-9d95-9c6586081175 {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 846.282533] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8baf3509-4d25-4d69-bd0d-4210316a1e20 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.288656] env[62460]: DEBUG oslo_vmware.api [None req-759824f9-86b1-41cc-87e9-b267e31f479b tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 846.288656] env[62460]: value = "task-1313635" [ 846.288656] env[62460]: _type = "Task" [ 846.288656] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.293664] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Releasing lock "refresh_cache-16d40829-ba6e-4193-98b5-fff94c066bcb" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.294090] env[62460]: DEBUG nova.compute.manager [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] Instance network_info: |[{"id": "b1e3ec03-310d-4b60-8ead-a10ec31a5584", "address": "fa:16:3e:7c:f9:2b", "network": {"id": "c8b3d824-c8f0-4f7b-a91f-1987e04a5f09", "bridge": "br-int", "label": "tempest-ServersTestJSON-1769082812-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10ac54db1ef54e249a077bbb3dde4242", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "255460d5-71d4-4bfd-87f1-acc10085db7f", "external-id": "nsx-vlan-transportzone-152", "segmentation_id": 152, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb1e3ec03-31", "ovs_interfaceid": "b1e3ec03-310d-4b60-8ead-a10ec31a5584", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 846.298029] env[62460]: DEBUG oslo_concurrency.lockutils [req-74938e89-a9d0-47b6-9b1a-86e2825bc273 req-1bd8eaeb-75dd-4bdf-ae42-1a30ee118627 service nova] Acquired lock "refresh_cache-16d40829-ba6e-4193-98b5-fff94c066bcb" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.298029] env[62460]: DEBUG nova.network.neutron [req-74938e89-a9d0-47b6-9b1a-86e2825bc273 req-1bd8eaeb-75dd-4bdf-ae42-1a30ee118627 service nova] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] Refreshing network info cache for port b1e3ec03-310d-4b60-8ead-a10ec31a5584 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 846.298699] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7c:f9:2b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '255460d5-71d4-4bfd-87f1-acc10085db7f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b1e3ec03-310d-4b60-8ead-a10ec31a5584', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 846.305940] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Creating folder: Project (10ac54db1ef54e249a077bbb3dde4242). Parent ref: group-v281134. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 846.306213] env[62460]: DEBUG oslo_vmware.api [None req-759824f9-86b1-41cc-87e9-b267e31f479b tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313635, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.307218] env[62460]: DEBUG nova.scheduler.client.report [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 846.310657] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-18d4b3dd-6f3a-4242-b305-7cfb62330e16 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.321105] env[62460]: DEBUG oslo_vmware.api [None req-af702ee7-e128-4206-aea5-9106143dcc05 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313628, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.323137] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Created folder: Project (10ac54db1ef54e249a077bbb3dde4242) in parent group-v281134. [ 846.323137] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Creating folder: Instances. Parent ref: group-v281197. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 846.323137] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-246ba070-852f-452a-9f77-33327cf074cb {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.330635] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Created folder: Instances in parent group-v281197. [ 846.330792] env[62460]: DEBUG oslo.service.loopingcall [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 846.330978] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 846.331192] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2f1c4b92-7460-4c0a-a1a0-0a67c1d27a8f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.349293] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 846.349293] env[62460]: value = "task-1313638" [ 846.349293] env[62460]: _type = "Task" [ 846.349293] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.357317] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313638, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.432935] env[62460]: DEBUG nova.network.neutron [-] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.436666] env[62460]: DEBUG nova.compute.manager [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 846.477522] env[62460]: DEBUG nova.virt.hardware [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 846.477809] env[62460]: DEBUG nova.virt.hardware [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 846.478032] env[62460]: DEBUG nova.virt.hardware [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 846.478349] env[62460]: DEBUG nova.virt.hardware [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 846.478575] env[62460]: DEBUG nova.virt.hardware [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 846.478806] env[62460]: DEBUG nova.virt.hardware [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 846.479107] env[62460]: DEBUG nova.virt.hardware [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 846.479357] env[62460]: DEBUG nova.virt.hardware [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 846.479619] env[62460]: DEBUG nova.virt.hardware [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 846.479885] env[62460]: DEBUG nova.virt.hardware [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 846.480168] env[62460]: DEBUG nova.virt.hardware [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 846.481174] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6862e5d2-43c3-48d5-bc44-cb66bdd990b1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.498773] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a67f4b60-b1f5-4242-afec-5dea512d76c1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.503658] env[62460]: DEBUG oslo_vmware.api [None req-53ff2f89-d24e-47c8-a414-09b17968e3cd tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313631, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.512629] env[62460]: DEBUG oslo_vmware.api [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313633, 'name': ReconfigVM_Task, 'duration_secs': 0.282624} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.522905] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Reconfigured VM instance instance-00000046 to attach disk [datastore2] d7d180f7-e1a8-46c8-ba8e-ca50dac474cc/d7d180f7-e1a8-46c8-ba8e-ca50dac474cc.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 846.523356] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8ba5388f-3ce8-460a-9ad9-e3056e0a35f9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.529330] env[62460]: DEBUG oslo_vmware.api [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 846.529330] env[62460]: value = "task-1313639" [ 846.529330] env[62460]: _type = "Task" [ 846.529330] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.538201] env[62460]: DEBUG oslo_vmware.api [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313639, 'name': Rename_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.649221] env[62460]: DEBUG nova.network.neutron [None req-a781a4a5-2201-450d-aedf-b182703d796e tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Successfully created port: bf0f715f-7988-45f9-81b6-c92688ff60c7 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 846.772010] env[62460]: DEBUG oslo_concurrency.lockutils [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "fd03509c-5471-48bb-8150-d29531dca848" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.772267] env[62460]: DEBUG oslo_concurrency.lockutils [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "fd03509c-5471-48bb-8150-d29531dca848" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.798167] env[62460]: DEBUG oslo_vmware.api [None req-759824f9-86b1-41cc-87e9-b267e31f479b tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313635, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.217943} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.798407] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-759824f9-86b1-41cc-87e9-b267e31f479b tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 846.798611] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-759824f9-86b1-41cc-87e9-b267e31f479b tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Deleted contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 846.798820] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-759824f9-86b1-41cc-87e9-b267e31f479b tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 846.799038] env[62460]: INFO nova.compute.manager [None req-759824f9-86b1-41cc-87e9-b267e31f479b tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Took 1.12 seconds to destroy the instance on the hypervisor. [ 846.799287] env[62460]: DEBUG oslo.service.loopingcall [None req-759824f9-86b1-41cc-87e9-b267e31f479b tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 846.799478] env[62460]: DEBUG nova.compute.manager [-] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 846.799574] env[62460]: DEBUG nova.network.neutron [-] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 846.813396] env[62460]: DEBUG oslo_concurrency.lockutils [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.404s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.813900] env[62460]: DEBUG nova.compute.manager [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 846.816409] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.384s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.817838] env[62460]: INFO nova.compute.claims [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 846.828923] env[62460]: DEBUG oslo_vmware.api [None req-af702ee7-e128-4206-aea5-9106143dcc05 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313628, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.861423] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313638, 'name': CreateVM_Task} progress is 6%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.935976] env[62460]: INFO nova.compute.manager [-] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] Took 1.47 seconds to deallocate network for instance. [ 846.996459] env[62460]: DEBUG oslo_vmware.api [None req-53ff2f89-d24e-47c8-a414-09b17968e3cd tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313631, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.048020] env[62460]: DEBUG oslo_vmware.api [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313639, 'name': Rename_Task, 'duration_secs': 0.131549} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.048020] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 847.048020] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d55b2ee8-e951-4453-8244-48b79e4fbbfb {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.052816] env[62460]: DEBUG oslo_vmware.api [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 847.052816] env[62460]: value = "task-1313640" [ 847.052816] env[62460]: _type = "Task" [ 847.052816] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.068827] env[62460]: DEBUG oslo_vmware.api [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313640, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.274485] env[62460]: DEBUG nova.compute.manager [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 847.284544] env[62460]: DEBUG nova.network.neutron [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Successfully updated port: c6936c73-cdb0-4475-b00b-de47fc29e7cd {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 847.319164] env[62460]: DEBUG nova.network.neutron [req-74938e89-a9d0-47b6-9b1a-86e2825bc273 req-1bd8eaeb-75dd-4bdf-ae42-1a30ee118627 service nova] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] Updated VIF entry in instance network info cache for port b1e3ec03-310d-4b60-8ead-a10ec31a5584. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 847.319164] env[62460]: DEBUG nova.network.neutron [req-74938e89-a9d0-47b6-9b1a-86e2825bc273 req-1bd8eaeb-75dd-4bdf-ae42-1a30ee118627 service nova] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] Updating instance_info_cache with network_info: [{"id": "b1e3ec03-310d-4b60-8ead-a10ec31a5584", "address": "fa:16:3e:7c:f9:2b", "network": {"id": "c8b3d824-c8f0-4f7b-a91f-1987e04a5f09", "bridge": "br-int", "label": "tempest-ServersTestJSON-1769082812-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10ac54db1ef54e249a077bbb3dde4242", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "255460d5-71d4-4bfd-87f1-acc10085db7f", "external-id": "nsx-vlan-transportzone-152", "segmentation_id": 152, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb1e3ec03-31", "ovs_interfaceid": "b1e3ec03-310d-4b60-8ead-a10ec31a5584", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.325969] env[62460]: DEBUG nova.compute.utils [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 847.328841] env[62460]: DEBUG nova.compute.manager [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 847.329093] env[62460]: DEBUG nova.network.neutron [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 847.338382] env[62460]: DEBUG oslo_vmware.api [None req-af702ee7-e128-4206-aea5-9106143dcc05 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313628, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.359470] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313638, 'name': CreateVM_Task} progress is 15%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.376933] env[62460]: DEBUG nova.policy [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f8cbad575ee04149a9b63d68559f5c66', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'af1a43a6e9824437952b920309572ab3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 847.443780] env[62460]: DEBUG oslo_concurrency.lockutils [None req-810313b0-6843-4505-a7de-48602dfc2f7b tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.456545] env[62460]: INFO nova.compute.manager [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Rebuilding instance [ 847.496795] env[62460]: DEBUG oslo_vmware.api [None req-53ff2f89-d24e-47c8-a414-09b17968e3cd tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313631, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.501029] env[62460]: DEBUG nova.compute.manager [req-b609ac41-f6d5-4131-9c77-d020e69c7cf3 req-c753bf31-92d6-4c0f-9ee7-8e6129ea8e3b service nova] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] Received event network-vif-deleted-91d45497-1367-4747-acd1-984bd67bc127 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 847.506824] env[62460]: DEBUG nova.compute.manager [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 847.507638] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3ccdd37-d794-4a79-b7e8-17dcda6572b0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.563484] env[62460]: DEBUG oslo_vmware.api [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313640, 'name': PowerOnVM_Task, 'duration_secs': 0.466054} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.563790] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 847.564098] env[62460]: INFO nova.compute.manager [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Took 5.94 seconds to spawn the instance on the hypervisor. [ 847.564306] env[62460]: DEBUG nova.compute.manager [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 847.565099] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f4245ea-ec23-40e9-80aa-8c2ad40d12b3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.636938] env[62460]: DEBUG nova.network.neutron [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] Successfully created port: 00526c55-f6d3-48d6-ba66-a1f821f2b03b {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 847.760917] env[62460]: DEBUG nova.compute.manager [req-beab4e26-a6d1-4525-a134-c6507af01586 req-3347c6dd-10c0-46a3-bcd5-9a39265b3a40 service nova] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Received event network-vif-deleted-bcdde0bd-b227-4477-ae73-bebbae418c5e {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 847.761131] env[62460]: INFO nova.compute.manager [req-beab4e26-a6d1-4525-a134-c6507af01586 req-3347c6dd-10c0-46a3-bcd5-9a39265b3a40 service nova] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Neutron deleted interface bcdde0bd-b227-4477-ae73-bebbae418c5e; detaching it from the instance and deleting it from the info cache [ 847.761307] env[62460]: DEBUG nova.network.neutron [req-beab4e26-a6d1-4525-a134-c6507af01586 req-3347c6dd-10c0-46a3-bcd5-9a39265b3a40 service nova] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.783837] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Acquiring lock "refresh_cache-a02b4be1-91b0-4254-8d60-654885e24f6b" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.783986] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Acquired lock "refresh_cache-a02b4be1-91b0-4254-8d60-654885e24f6b" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.784153] env[62460]: DEBUG nova.network.neutron [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 847.796388] env[62460]: DEBUG oslo_concurrency.lockutils [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.824664] env[62460]: DEBUG oslo_concurrency.lockutils [req-74938e89-a9d0-47b6-9b1a-86e2825bc273 req-1bd8eaeb-75dd-4bdf-ae42-1a30ee118627 service nova] Releasing lock "refresh_cache-16d40829-ba6e-4193-98b5-fff94c066bcb" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.831984] env[62460]: DEBUG nova.compute.manager [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 847.833632] env[62460]: DEBUG oslo_vmware.api [None req-af702ee7-e128-4206-aea5-9106143dcc05 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313628, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.847515] env[62460]: DEBUG nova.network.neutron [-] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.865446] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313638, 'name': CreateVM_Task} progress is 15%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.998127] env[62460]: DEBUG oslo_vmware.api [None req-53ff2f89-d24e-47c8-a414-09b17968e3cd tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313631, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.018715] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 848.019434] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-520dccda-084f-4a6e-9225-96f23d3a43d2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.029932] env[62460]: DEBUG oslo_vmware.api [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Waiting for the task: (returnval){ [ 848.029932] env[62460]: value = "task-1313641" [ 848.029932] env[62460]: _type = "Task" [ 848.029932] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.037265] env[62460]: DEBUG oslo_vmware.api [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Task: {'id': task-1313641, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.099884] env[62460]: INFO nova.compute.manager [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Took 28.97 seconds to build instance. [ 848.138225] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8814e468-0655-4387-bdf8-d122718002f0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.144470] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-beaf97a3-ea96-44a9-8349-f461175fb74f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.183735] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8d6b9d7-eedb-48f7-a434-06eedebc0e1b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.192236] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64ee61d8-7024-4bed-a24e-78fad2136058 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.207757] env[62460]: DEBUG nova.compute.provider_tree [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 848.264346] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-54340b54-7396-4bb4-bafe-df9a117381ef {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.276727] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-926616b8-e448-41d9-a162-8c73197ff203 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.308577] env[62460]: DEBUG nova.compute.manager [req-beab4e26-a6d1-4525-a134-c6507af01586 req-3347c6dd-10c0-46a3-bcd5-9a39265b3a40 service nova] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Detach interface failed, port_id=bcdde0bd-b227-4477-ae73-bebbae418c5e, reason: Instance b76f37a0-91d0-4a01-9d95-9c6586081175 could not be found. {{(pid=62460) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 848.318737] env[62460]: DEBUG nova.network.neutron [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 848.330673] env[62460]: DEBUG oslo_vmware.api [None req-af702ee7-e128-4206-aea5-9106143dcc05 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313628, 'name': DeleteDatastoreFile_Task, 'duration_secs': 2.779208} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.331015] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-af702ee7-e128-4206-aea5-9106143dcc05 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 848.331216] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-af702ee7-e128-4206-aea5-9106143dcc05 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] Deleted contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 848.331396] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-af702ee7-e128-4206-aea5-9106143dcc05 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 848.331577] env[62460]: INFO nova.compute.manager [None req-af702ee7-e128-4206-aea5-9106143dcc05 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] Took 3.79 seconds to destroy the instance on the hypervisor. [ 848.331825] env[62460]: DEBUG oslo.service.loopingcall [None req-af702ee7-e128-4206-aea5-9106143dcc05 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 848.332061] env[62460]: DEBUG nova.compute.manager [-] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 848.332214] env[62460]: DEBUG nova.network.neutron [-] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 848.361728] env[62460]: INFO nova.compute.manager [-] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Took 1.56 seconds to deallocate network for instance. [ 848.361978] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313638, 'name': CreateVM_Task, 'duration_secs': 1.979148} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.363549] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 848.364158] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 848.364325] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.364647] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 848.367439] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-99ea6f1c-320b-4b32-ac7c-99c7b5c651a3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.373014] env[62460]: DEBUG nova.network.neutron [None req-a781a4a5-2201-450d-aedf-b182703d796e tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Successfully updated port: bf0f715f-7988-45f9-81b6-c92688ff60c7 {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 848.375556] env[62460]: DEBUG oslo_vmware.api [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for the task: (returnval){ [ 848.375556] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52d2b728-bf36-23c8-a679-63b4e8e65332" [ 848.375556] env[62460]: _type = "Task" [ 848.375556] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.384811] env[62460]: DEBUG oslo_vmware.api [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52d2b728-bf36-23c8-a679-63b4e8e65332, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.482472] env[62460]: DEBUG nova.network.neutron [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Updating instance_info_cache with network_info: [{"id": "c6936c73-cdb0-4475-b00b-de47fc29e7cd", "address": "fa:16:3e:4b:3f:85", "network": {"id": "25cd6407-cd9f-4e8b-8dd6-0c585bb26760", "bridge": "br-int", "label": "tempest-ImagesTestJSON-175584043-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "11491a5610734d96bac82035c46bd5d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24ec44b7-0acf-4ff9-8bb3-4641b74af7a7", "external-id": "nsx-vlan-transportzone-705", "segmentation_id": 705, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6936c73-cd", "ovs_interfaceid": "c6936c73-cdb0-4475-b00b-de47fc29e7cd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.499739] env[62460]: DEBUG oslo_vmware.api [None req-53ff2f89-d24e-47c8-a414-09b17968e3cd tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313631, 'name': DeleteDatastoreFile_Task, 'duration_secs': 2.615218} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.499933] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-53ff2f89-d24e-47c8-a414-09b17968e3cd tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 848.500258] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-53ff2f89-d24e-47c8-a414-09b17968e3cd tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] Deleted contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 848.500524] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-53ff2f89-d24e-47c8-a414-09b17968e3cd tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 848.500716] env[62460]: INFO nova.compute.manager [None req-53ff2f89-d24e-47c8-a414-09b17968e3cd tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] Took 3.65 seconds to destroy the instance on the hypervisor. [ 848.500962] env[62460]: DEBUG oslo.service.loopingcall [None req-53ff2f89-d24e-47c8-a414-09b17968e3cd tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 848.501166] env[62460]: DEBUG nova.compute.manager [-] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 848.501258] env[62460]: DEBUG nova.network.neutron [-] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 848.538520] env[62460]: DEBUG oslo_vmware.api [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Task: {'id': task-1313641, 'name': PowerOffVM_Task, 'duration_secs': 0.162645} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.539232] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 848.539726] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 848.540483] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bd2c3ed-97a2-478d-aeb3-5bcbcb7e1758 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.547685] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 848.547924] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a465a526-6d69-457f-9564-10eafb3a4031 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.573132] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 848.573510] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Deleting contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 848.573593] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Deleting the datastore file [datastore2] 9d1a5830-f3c0-4d18-9338-16f7b6962c6a {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 848.573876] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d414af16-4c16-4f10-a6ba-bb7f715a01b4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.580323] env[62460]: DEBUG oslo_vmware.api [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Waiting for the task: (returnval){ [ 848.580323] env[62460]: value = "task-1313643" [ 848.580323] env[62460]: _type = "Task" [ 848.580323] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.591279] env[62460]: DEBUG oslo_vmware.api [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Task: {'id': task-1313643, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.602945] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1dafecf1-5ea7-400f-91bf-c4ee9a7462df tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "d7d180f7-e1a8-46c8-ba8e-ca50dac474cc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 84.062s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.711194] env[62460]: DEBUG nova.scheduler.client.report [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 848.839115] env[62460]: DEBUG nova.compute.manager [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 848.864358] env[62460]: DEBUG nova.virt.hardware [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 848.864634] env[62460]: DEBUG nova.virt.hardware [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 848.864799] env[62460]: DEBUG nova.virt.hardware [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 848.864982] env[62460]: DEBUG nova.virt.hardware [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 848.865148] env[62460]: DEBUG nova.virt.hardware [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 848.865298] env[62460]: DEBUG nova.virt.hardware [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 848.865509] env[62460]: DEBUG nova.virt.hardware [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 848.865667] env[62460]: DEBUG nova.virt.hardware [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 848.865864] env[62460]: DEBUG nova.virt.hardware [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 848.866049] env[62460]: DEBUG nova.virt.hardware [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 848.866228] env[62460]: DEBUG nova.virt.hardware [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 848.867459] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe91ccc1-07eb-43be-b5e7-031954c0f95d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.874498] env[62460]: DEBUG oslo_concurrency.lockutils [None req-759824f9-86b1-41cc-87e9-b267e31f479b tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.875799] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0834830-2025-4e84-8c6c-ca773952c8ce {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.881394] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a781a4a5-2201-450d-aedf-b182703d796e tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "refresh_cache-7e6ff902-4a04-43d5-9014-38c4ec88efc4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 848.881590] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a781a4a5-2201-450d-aedf-b182703d796e tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquired lock "refresh_cache-7e6ff902-4a04-43d5-9014-38c4ec88efc4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.881745] env[62460]: DEBUG nova.network.neutron [None req-a781a4a5-2201-450d-aedf-b182703d796e tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 848.899993] env[62460]: DEBUG oslo_vmware.api [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52d2b728-bf36-23c8-a679-63b4e8e65332, 'name': SearchDatastore_Task, 'duration_secs': 0.014726} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.900323] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.900549] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 848.900777] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 848.900922] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.901130] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 848.901644] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8a6c5233-c55b-4051-b4bc-8e36cc4a570b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.910408] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 848.910539] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 848.912010] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fd8521ed-ce3b-443d-bb85-795e9a81452a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.916514] env[62460]: DEBUG oslo_vmware.api [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for the task: (returnval){ [ 848.916514] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]525ec0e2-f61a-0ab7-8836-dce643f42875" [ 848.916514] env[62460]: _type = "Task" [ 848.916514] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.923879] env[62460]: DEBUG oslo_vmware.api [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]525ec0e2-f61a-0ab7-8836-dce643f42875, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.984864] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Releasing lock "refresh_cache-a02b4be1-91b0-4254-8d60-654885e24f6b" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.985271] env[62460]: DEBUG nova.compute.manager [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Instance network_info: |[{"id": "c6936c73-cdb0-4475-b00b-de47fc29e7cd", "address": "fa:16:3e:4b:3f:85", "network": {"id": "25cd6407-cd9f-4e8b-8dd6-0c585bb26760", "bridge": "br-int", "label": "tempest-ImagesTestJSON-175584043-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "11491a5610734d96bac82035c46bd5d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24ec44b7-0acf-4ff9-8bb3-4641b74af7a7", "external-id": "nsx-vlan-transportzone-705", "segmentation_id": 705, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6936c73-cd", "ovs_interfaceid": "c6936c73-cdb0-4475-b00b-de47fc29e7cd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 848.985694] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4b:3f:85', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '24ec44b7-0acf-4ff9-8bb3-4641b74af7a7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c6936c73-cdb0-4475-b00b-de47fc29e7cd', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 848.993621] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Creating folder: Project (11491a5610734d96bac82035c46bd5d1). Parent ref: group-v281134. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 848.994678] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-14ca99f9-3888-4b0a-b7f0-b05e2aa84c66 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.006313] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Created folder: Project (11491a5610734d96bac82035c46bd5d1) in parent group-v281134. [ 849.006313] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Creating folder: Instances. Parent ref: group-v281200. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 849.006313] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-775e7942-dd2b-4062-812c-e10d96964c76 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.016423] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Created folder: Instances in parent group-v281200. [ 849.016727] env[62460]: DEBUG oslo.service.loopingcall [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 849.016970] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 849.017239] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d544c06b-342f-40f1-bd6b-9b3e36605cc1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.037090] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 849.037090] env[62460]: value = "task-1313646" [ 849.037090] env[62460]: _type = "Task" [ 849.037090] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.046440] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313646, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.090396] env[62460]: DEBUG oslo_vmware.api [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Task: {'id': task-1313643, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.320857} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.090816] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 849.090921] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Deleted contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 849.091874] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 849.111459] env[62460]: DEBUG nova.network.neutron [-] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 849.153320] env[62460]: DEBUG nova.network.neutron [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] Successfully updated port: 00526c55-f6d3-48d6-ba66-a1f821f2b03b {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 849.216026] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.399s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.216904] env[62460]: DEBUG nova.compute.manager [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 849.219283] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c2624ccd-5dd5-44c5-9618-3e97c3395a8c tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.116s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.219560] env[62460]: DEBUG nova.objects.instance [None req-c2624ccd-5dd5-44c5-9618-3e97c3395a8c tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Lazy-loading 'resources' on Instance uuid 13890351-6091-4b4f-8484-1cdd0c8523b0 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 849.228068] env[62460]: DEBUG nova.network.neutron [-] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 849.427024] env[62460]: DEBUG oslo_vmware.api [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]525ec0e2-f61a-0ab7-8836-dce643f42875, 'name': SearchDatastore_Task, 'duration_secs': 0.029498} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.427896] env[62460]: WARNING nova.network.neutron [None req-a781a4a5-2201-450d-aedf-b182703d796e tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] d3136e32-ad55-4b73-835c-8fa5f0480767 already exists in list: networks containing: ['d3136e32-ad55-4b73-835c-8fa5f0480767']. ignoring it [ 849.430155] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f5841cb2-dec3-4d0e-b772-cf9e6cecf0f0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.435143] env[62460]: DEBUG oslo_vmware.api [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for the task: (returnval){ [ 849.435143] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]529cb09b-9ee9-6288-035c-d9100e889c3e" [ 849.435143] env[62460]: _type = "Task" [ 849.435143] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.444937] env[62460]: DEBUG oslo_vmware.api [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]529cb09b-9ee9-6288-035c-d9100e889c3e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.452424] env[62460]: DEBUG oslo_concurrency.lockutils [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "d7d180f7-e1a8-46c8-ba8e-ca50dac474cc" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.452659] env[62460]: DEBUG oslo_concurrency.lockutils [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "d7d180f7-e1a8-46c8-ba8e-ca50dac474cc" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.452843] env[62460]: INFO nova.compute.manager [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Shelving [ 849.546783] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313646, 'name': CreateVM_Task} progress is 99%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.614603] env[62460]: INFO nova.compute.manager [-] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] Took 1.28 seconds to deallocate network for instance. [ 849.656361] env[62460]: DEBUG oslo_concurrency.lockutils [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Acquiring lock "refresh_cache-49477d35-92ea-4f9d-8333-fc23144b7dfc" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.657091] env[62460]: DEBUG oslo_concurrency.lockutils [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Acquired lock "refresh_cache-49477d35-92ea-4f9d-8333-fc23144b7dfc" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.657091] env[62460]: DEBUG nova.network.neutron [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 849.725231] env[62460]: DEBUG nova.compute.utils [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 849.726829] env[62460]: DEBUG nova.compute.manager [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 849.726996] env[62460]: DEBUG nova.network.neutron [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 849.730265] env[62460]: INFO nova.compute.manager [-] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] Took 1.23 seconds to deallocate network for instance. [ 849.754173] env[62460]: DEBUG nova.network.neutron [None req-a781a4a5-2201-450d-aedf-b182703d796e tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Updating instance_info_cache with network_info: [{"id": "ba38cbb0-62cf-47cb-a238-87ebc03de250", "address": "fa:16:3e:b4:93:18", "network": {"id": "d3136e32-ad55-4b73-835c-8fa5f0480767", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1511414968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53fa06008e9f43488362895e7a143700", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "054fcd1e-638e-425a-a1de-78cb188ae026", "external-id": "nsx-vlan-transportzone-658", "segmentation_id": 658, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba38cbb0-62", "ovs_interfaceid": "ba38cbb0-62cf-47cb-a238-87ebc03de250", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "bf0f715f-7988-45f9-81b6-c92688ff60c7", "address": "fa:16:3e:84:3c:55", "network": {"id": "d3136e32-ad55-4b73-835c-8fa5f0480767", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1511414968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53fa06008e9f43488362895e7a143700", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "054fcd1e-638e-425a-a1de-78cb188ae026", "external-id": "nsx-vlan-transportzone-658", "segmentation_id": 658, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf0f715f-79", "ovs_interfaceid": "bf0f715f-7988-45f9-81b6-c92688ff60c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 849.775622] env[62460]: DEBUG nova.policy [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '947371ed3ed94ee5b2a900d47444791c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f3d2d1c48ec14121a2e8c9b3f800a949', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 849.952406] env[62460]: DEBUG oslo_vmware.api [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]529cb09b-9ee9-6288-035c-d9100e889c3e, 'name': SearchDatastore_Task, 'duration_secs': 0.010802} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.952626] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.952911] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 16d40829-ba6e-4193-98b5-fff94c066bcb/16d40829-ba6e-4193-98b5-fff94c066bcb.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 849.953203] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d48f20ee-e76b-4f21-a84d-5e320b82fc30 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.962330] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 849.962582] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0068ac91-c00a-46b7-bff1-4fc86aec14ec {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.965804] env[62460]: DEBUG oslo_vmware.api [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for the task: (returnval){ [ 849.965804] env[62460]: value = "task-1313647" [ 849.965804] env[62460]: _type = "Task" [ 849.965804] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.972787] env[62460]: DEBUG oslo_vmware.api [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 849.972787] env[62460]: value = "task-1313648" [ 849.972787] env[62460]: _type = "Task" [ 849.972787] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.979142] env[62460]: DEBUG oslo_vmware.api [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313647, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.983840] env[62460]: DEBUG oslo_vmware.api [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313648, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.998888] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9bb5050-d79e-4e21-ab23-ebb2b497a9a6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.006445] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea3c90a7-a9b5-4d0d-b0c4-01bde84952c6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.043450] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-345fc6c9-fb0b-44f3-bf18-b8163e8964af {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.046635] env[62460]: DEBUG nova.network.neutron [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Successfully created port: 72a87ddf-0585-429a-b9de-d73bcad42cd1 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 850.052959] env[62460]: DEBUG nova.compute.manager [req-7cc1ff3e-848a-4301-ba7a-e0c4a0a11cf1 req-9ff1a9c2-f3b6-45c3-b908-61470d180643 service nova] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Received event network-vif-plugged-c6936c73-cdb0-4475-b00b-de47fc29e7cd {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 850.053187] env[62460]: DEBUG oslo_concurrency.lockutils [req-7cc1ff3e-848a-4301-ba7a-e0c4a0a11cf1 req-9ff1a9c2-f3b6-45c3-b908-61470d180643 service nova] Acquiring lock "a02b4be1-91b0-4254-8d60-654885e24f6b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.053395] env[62460]: DEBUG oslo_concurrency.lockutils [req-7cc1ff3e-848a-4301-ba7a-e0c4a0a11cf1 req-9ff1a9c2-f3b6-45c3-b908-61470d180643 service nova] Lock "a02b4be1-91b0-4254-8d60-654885e24f6b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.053619] env[62460]: DEBUG oslo_concurrency.lockutils [req-7cc1ff3e-848a-4301-ba7a-e0c4a0a11cf1 req-9ff1a9c2-f3b6-45c3-b908-61470d180643 service nova] Lock "a02b4be1-91b0-4254-8d60-654885e24f6b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.053877] env[62460]: DEBUG nova.compute.manager [req-7cc1ff3e-848a-4301-ba7a-e0c4a0a11cf1 req-9ff1a9c2-f3b6-45c3-b908-61470d180643 service nova] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] No waiting events found dispatching network-vif-plugged-c6936c73-cdb0-4475-b00b-de47fc29e7cd {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 850.054022] env[62460]: WARNING nova.compute.manager [req-7cc1ff3e-848a-4301-ba7a-e0c4a0a11cf1 req-9ff1a9c2-f3b6-45c3-b908-61470d180643 service nova] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Received unexpected event network-vif-plugged-c6936c73-cdb0-4475-b00b-de47fc29e7cd for instance with vm_state building and task_state spawning. [ 850.054160] env[62460]: DEBUG nova.compute.manager [req-7cc1ff3e-848a-4301-ba7a-e0c4a0a11cf1 req-9ff1a9c2-f3b6-45c3-b908-61470d180643 service nova] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Received event network-changed-c6936c73-cdb0-4475-b00b-de47fc29e7cd {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 850.054318] env[62460]: DEBUG nova.compute.manager [req-7cc1ff3e-848a-4301-ba7a-e0c4a0a11cf1 req-9ff1a9c2-f3b6-45c3-b908-61470d180643 service nova] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Refreshing instance network info cache due to event network-changed-c6936c73-cdb0-4475-b00b-de47fc29e7cd. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 850.054502] env[62460]: DEBUG oslo_concurrency.lockutils [req-7cc1ff3e-848a-4301-ba7a-e0c4a0a11cf1 req-9ff1a9c2-f3b6-45c3-b908-61470d180643 service nova] Acquiring lock "refresh_cache-a02b4be1-91b0-4254-8d60-654885e24f6b" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.054637] env[62460]: DEBUG oslo_concurrency.lockutils [req-7cc1ff3e-848a-4301-ba7a-e0c4a0a11cf1 req-9ff1a9c2-f3b6-45c3-b908-61470d180643 service nova] Acquired lock "refresh_cache-a02b4be1-91b0-4254-8d60-654885e24f6b" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.054790] env[62460]: DEBUG nova.network.neutron [req-7cc1ff3e-848a-4301-ba7a-e0c4a0a11cf1 req-9ff1a9c2-f3b6-45c3-b908-61470d180643 service nova] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Refreshing network info cache for port c6936c73-cdb0-4475-b00b-de47fc29e7cd {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 850.061127] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313646, 'name': CreateVM_Task, 'duration_secs': 0.5943} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.062611] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f35bf7f7-c29d-4f33-8556-d9ed99afabf0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.066923] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 850.068179] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.068359] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.068697] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 850.069604] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c391645d-712b-4651-ba38-2f539031578c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.080198] env[62460]: DEBUG nova.compute.provider_tree [None req-c2624ccd-5dd5-44c5-9618-3e97c3395a8c tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 850.085179] env[62460]: DEBUG oslo_vmware.api [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Waiting for the task: (returnval){ [ 850.085179] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52a0b7b8-2556-6814-0d73-a41d4d3ebd6e" [ 850.085179] env[62460]: _type = "Task" [ 850.085179] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.094881] env[62460]: DEBUG oslo_vmware.api [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52a0b7b8-2556-6814-0d73-a41d4d3ebd6e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.125114] env[62460]: DEBUG oslo_concurrency.lockutils [None req-af702ee7-e128-4206-aea5-9106143dcc05 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.136128] env[62460]: DEBUG nova.virt.hardware [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 850.136411] env[62460]: DEBUG nova.virt.hardware [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 850.136572] env[62460]: DEBUG nova.virt.hardware [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 850.136766] env[62460]: DEBUG nova.virt.hardware [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 850.136928] env[62460]: DEBUG nova.virt.hardware [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 850.137094] env[62460]: DEBUG nova.virt.hardware [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 850.140348] env[62460]: DEBUG nova.virt.hardware [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 850.140348] env[62460]: DEBUG nova.virt.hardware [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 850.140348] env[62460]: DEBUG nova.virt.hardware [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 850.140348] env[62460]: DEBUG nova.virt.hardware [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 850.140348] env[62460]: DEBUG nova.virt.hardware [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 850.140348] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b36e5c7-e27e-4e0f-ae7d-adedabdc9a38 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.147693] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-802d8ae8-8852-4111-a7a7-4034f2ded26e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.165850] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Instance VIF info [] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 850.171950] env[62460]: DEBUG oslo.service.loopingcall [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 850.172305] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 850.172940] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-58957e8a-46c4-461e-9131-5542e548e5c5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.187561] env[62460]: DEBUG nova.compute.manager [req-ad8f3779-3930-475e-83d8-67485ec0063a req-eb8e43bc-d379-4b38-bfe1-3082108484b2 service nova] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] Received event network-vif-deleted-42bbd4ac-f793-4c77-96c8-d45f34b8bb15 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 850.187775] env[62460]: DEBUG nova.compute.manager [req-ad8f3779-3930-475e-83d8-67485ec0063a req-eb8e43bc-d379-4b38-bfe1-3082108484b2 service nova] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] Received event network-vif-plugged-00526c55-f6d3-48d6-ba66-a1f821f2b03b {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 850.187980] env[62460]: DEBUG oslo_concurrency.lockutils [req-ad8f3779-3930-475e-83d8-67485ec0063a req-eb8e43bc-d379-4b38-bfe1-3082108484b2 service nova] Acquiring lock "49477d35-92ea-4f9d-8333-fc23144b7dfc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.188193] env[62460]: DEBUG oslo_concurrency.lockutils [req-ad8f3779-3930-475e-83d8-67485ec0063a req-eb8e43bc-d379-4b38-bfe1-3082108484b2 service nova] Lock "49477d35-92ea-4f9d-8333-fc23144b7dfc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.188376] env[62460]: DEBUG oslo_concurrency.lockutils [req-ad8f3779-3930-475e-83d8-67485ec0063a req-eb8e43bc-d379-4b38-bfe1-3082108484b2 service nova] Lock "49477d35-92ea-4f9d-8333-fc23144b7dfc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.188546] env[62460]: DEBUG nova.compute.manager [req-ad8f3779-3930-475e-83d8-67485ec0063a req-eb8e43bc-d379-4b38-bfe1-3082108484b2 service nova] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] No waiting events found dispatching network-vif-plugged-00526c55-f6d3-48d6-ba66-a1f821f2b03b {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 850.188708] env[62460]: WARNING nova.compute.manager [req-ad8f3779-3930-475e-83d8-67485ec0063a req-eb8e43bc-d379-4b38-bfe1-3082108484b2 service nova] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] Received unexpected event network-vif-plugged-00526c55-f6d3-48d6-ba66-a1f821f2b03b for instance with vm_state building and task_state spawning. [ 850.188867] env[62460]: DEBUG nova.compute.manager [req-ad8f3779-3930-475e-83d8-67485ec0063a req-eb8e43bc-d379-4b38-bfe1-3082108484b2 service nova] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] Received event network-changed-00526c55-f6d3-48d6-ba66-a1f821f2b03b {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 850.189055] env[62460]: DEBUG nova.compute.manager [req-ad8f3779-3930-475e-83d8-67485ec0063a req-eb8e43bc-d379-4b38-bfe1-3082108484b2 service nova] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] Refreshing instance network info cache due to event network-changed-00526c55-f6d3-48d6-ba66-a1f821f2b03b. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 850.189221] env[62460]: DEBUG oslo_concurrency.lockutils [req-ad8f3779-3930-475e-83d8-67485ec0063a req-eb8e43bc-d379-4b38-bfe1-3082108484b2 service nova] Acquiring lock "refresh_cache-49477d35-92ea-4f9d-8333-fc23144b7dfc" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.195263] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 850.195263] env[62460]: value = "task-1313649" [ 850.195263] env[62460]: _type = "Task" [ 850.195263] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.203730] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313649, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.216055] env[62460]: DEBUG nova.network.neutron [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 850.230827] env[62460]: DEBUG nova.compute.manager [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 850.236235] env[62460]: DEBUG oslo_concurrency.lockutils [None req-53ff2f89-d24e-47c8-a414-09b17968e3cd tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.257312] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a781a4a5-2201-450d-aedf-b182703d796e tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Releasing lock "refresh_cache-7e6ff902-4a04-43d5-9014-38c4ec88efc4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.258033] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a781a4a5-2201-450d-aedf-b182703d796e tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "7e6ff902-4a04-43d5-9014-38c4ec88efc4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.258273] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a781a4a5-2201-450d-aedf-b182703d796e tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquired lock "7e6ff902-4a04-43d5-9014-38c4ec88efc4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.259552] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-464bf0e8-19fe-43a9-be00-c9031fccd401 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.280084] env[62460]: DEBUG nova.virt.hardware [None req-a781a4a5-2201-450d-aedf-b182703d796e tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 850.280442] env[62460]: DEBUG nova.virt.hardware [None req-a781a4a5-2201-450d-aedf-b182703d796e tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 850.280543] env[62460]: DEBUG nova.virt.hardware [None req-a781a4a5-2201-450d-aedf-b182703d796e tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 850.280796] env[62460]: DEBUG nova.virt.hardware [None req-a781a4a5-2201-450d-aedf-b182703d796e tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 850.280860] env[62460]: DEBUG nova.virt.hardware [None req-a781a4a5-2201-450d-aedf-b182703d796e tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 850.281141] env[62460]: DEBUG nova.virt.hardware [None req-a781a4a5-2201-450d-aedf-b182703d796e tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 850.281340] env[62460]: DEBUG nova.virt.hardware [None req-a781a4a5-2201-450d-aedf-b182703d796e tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 850.281573] env[62460]: DEBUG nova.virt.hardware [None req-a781a4a5-2201-450d-aedf-b182703d796e tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 850.281926] env[62460]: DEBUG nova.virt.hardware [None req-a781a4a5-2201-450d-aedf-b182703d796e tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 850.281926] env[62460]: DEBUG nova.virt.hardware [None req-a781a4a5-2201-450d-aedf-b182703d796e tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 850.285077] env[62460]: DEBUG nova.virt.hardware [None req-a781a4a5-2201-450d-aedf-b182703d796e tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 850.288950] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a781a4a5-2201-450d-aedf-b182703d796e tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Reconfiguring VM to attach interface {{(pid=62460) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 850.289353] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-74e86273-1d64-496b-816f-0b0edb23f1a7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.311689] env[62460]: DEBUG oslo_vmware.api [None req-a781a4a5-2201-450d-aedf-b182703d796e tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Waiting for the task: (returnval){ [ 850.311689] env[62460]: value = "task-1313650" [ 850.311689] env[62460]: _type = "Task" [ 850.311689] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.322467] env[62460]: DEBUG oslo_vmware.api [None req-a781a4a5-2201-450d-aedf-b182703d796e tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313650, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.477075] env[62460]: DEBUG oslo_vmware.api [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313647, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.478244] env[62460]: DEBUG nova.network.neutron [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] Updating instance_info_cache with network_info: [{"id": "00526c55-f6d3-48d6-ba66-a1f821f2b03b", "address": "fa:16:3e:33:a1:71", "network": {"id": "eea7af10-80a3-4d4a-b6d7-0ba4c1780cdf", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-719305021-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "af1a43a6e9824437952b920309572ab3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93341b73-918c-4e9d-9c66-ca171a54b574", "external-id": "nsx-vlan-transportzone-663", "segmentation_id": 663, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap00526c55-f6", "ovs_interfaceid": "00526c55-f6d3-48d6-ba66-a1f821f2b03b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.489009] env[62460]: DEBUG oslo_vmware.api [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313648, 'name': PowerOffVM_Task, 'duration_secs': 0.223944} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.489298] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 850.490895] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ed9a902-0428-4897-adff-f7646a228022 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.512670] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9689c6ae-a2e0-484a-9a7d-62a9473b8f81 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.586133] env[62460]: DEBUG nova.scheduler.client.report [None req-c2624ccd-5dd5-44c5-9618-3e97c3395a8c tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 850.600196] env[62460]: DEBUG oslo_vmware.api [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52a0b7b8-2556-6814-0d73-a41d4d3ebd6e, 'name': SearchDatastore_Task, 'duration_secs': 0.010135} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.600494] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.600726] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 850.601430] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.601430] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.601430] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 850.601567] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-002a011b-a90c-4cf9-b438-0820c0673ae2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.618847] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 850.619931] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 850.619931] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e490e9d5-ce3a-4330-b17d-954aa14ca02c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.625117] env[62460]: DEBUG oslo_vmware.api [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Waiting for the task: (returnval){ [ 850.625117] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52586062-8e6a-33b1-73e6-6c017285a045" [ 850.625117] env[62460]: _type = "Task" [ 850.625117] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.633889] env[62460]: DEBUG oslo_vmware.api [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52586062-8e6a-33b1-73e6-6c017285a045, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.706582] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313649, 'name': CreateVM_Task, 'duration_secs': 0.333602} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.706789] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 850.707257] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.707434] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.707765] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 850.708027] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b6af13b2-f410-4445-b319-5ed1c0e352fc {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.713082] env[62460]: DEBUG oslo_vmware.api [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Waiting for the task: (returnval){ [ 850.713082] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52d698ab-7e57-3dd4-c2a1-8cbb0b07683c" [ 850.713082] env[62460]: _type = "Task" [ 850.713082] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.721089] env[62460]: DEBUG oslo_vmware.api [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52d698ab-7e57-3dd4-c2a1-8cbb0b07683c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.771333] env[62460]: DEBUG nova.network.neutron [req-7cc1ff3e-848a-4301-ba7a-e0c4a0a11cf1 req-9ff1a9c2-f3b6-45c3-b908-61470d180643 service nova] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Updated VIF entry in instance network info cache for port c6936c73-cdb0-4475-b00b-de47fc29e7cd. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 850.771695] env[62460]: DEBUG nova.network.neutron [req-7cc1ff3e-848a-4301-ba7a-e0c4a0a11cf1 req-9ff1a9c2-f3b6-45c3-b908-61470d180643 service nova] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Updating instance_info_cache with network_info: [{"id": "c6936c73-cdb0-4475-b00b-de47fc29e7cd", "address": "fa:16:3e:4b:3f:85", "network": {"id": "25cd6407-cd9f-4e8b-8dd6-0c585bb26760", "bridge": "br-int", "label": "tempest-ImagesTestJSON-175584043-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "11491a5610734d96bac82035c46bd5d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24ec44b7-0acf-4ff9-8bb3-4641b74af7a7", "external-id": "nsx-vlan-transportzone-705", "segmentation_id": 705, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6936c73-cd", "ovs_interfaceid": "c6936c73-cdb0-4475-b00b-de47fc29e7cd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.821754] env[62460]: DEBUG oslo_vmware.api [None req-a781a4a5-2201-450d-aedf-b182703d796e tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313650, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.980122] env[62460]: DEBUG oslo_vmware.api [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313647, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.590004} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.980463] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 16d40829-ba6e-4193-98b5-fff94c066bcb/16d40829-ba6e-4193-98b5-fff94c066bcb.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 850.980714] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 850.981365] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fc5e800e-7b3e-4ced-8339-3c2bd43a9658 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.983202] env[62460]: DEBUG oslo_concurrency.lockutils [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Releasing lock "refresh_cache-49477d35-92ea-4f9d-8333-fc23144b7dfc" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.983488] env[62460]: DEBUG nova.compute.manager [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] Instance network_info: |[{"id": "00526c55-f6d3-48d6-ba66-a1f821f2b03b", "address": "fa:16:3e:33:a1:71", "network": {"id": "eea7af10-80a3-4d4a-b6d7-0ba4c1780cdf", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-719305021-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "af1a43a6e9824437952b920309572ab3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93341b73-918c-4e9d-9c66-ca171a54b574", "external-id": "nsx-vlan-transportzone-663", "segmentation_id": 663, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap00526c55-f6", "ovs_interfaceid": "00526c55-f6d3-48d6-ba66-a1f821f2b03b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 850.983767] env[62460]: DEBUG oslo_concurrency.lockutils [req-ad8f3779-3930-475e-83d8-67485ec0063a req-eb8e43bc-d379-4b38-bfe1-3082108484b2 service nova] Acquired lock "refresh_cache-49477d35-92ea-4f9d-8333-fc23144b7dfc" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.983956] env[62460]: DEBUG nova.network.neutron [req-ad8f3779-3930-475e-83d8-67485ec0063a req-eb8e43bc-d379-4b38-bfe1-3082108484b2 service nova] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] Refreshing network info cache for port 00526c55-f6d3-48d6-ba66-a1f821f2b03b {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 850.985257] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:33:a1:71', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '93341b73-918c-4e9d-9c66-ca171a54b574', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '00526c55-f6d3-48d6-ba66-a1f821f2b03b', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 850.993342] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Creating folder: Project (af1a43a6e9824437952b920309572ab3). Parent ref: group-v281134. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 850.994211] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cc24fc89-7484-41ea-8bbe-e512a9537b44 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.999330] env[62460]: DEBUG oslo_vmware.api [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for the task: (returnval){ [ 850.999330] env[62460]: value = "task-1313651" [ 850.999330] env[62460]: _type = "Task" [ 850.999330] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.005419] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Created folder: Project (af1a43a6e9824437952b920309572ab3) in parent group-v281134. [ 851.005419] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Creating folder: Instances. Parent ref: group-v281204. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 851.007948] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5f47e114-e60f-4d05-864d-66403b6ef7c6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.010149] env[62460]: DEBUG oslo_vmware.api [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313651, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.017348] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Created folder: Instances in parent group-v281204. [ 851.017600] env[62460]: DEBUG oslo.service.loopingcall [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 851.017802] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 851.018030] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4092a671-f749-47a4-8db1-424af5f39a53 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.034698] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Creating Snapshot of the VM instance {{(pid=62460) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 851.035146] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-55fcaef9-087e-4a0c-872f-283afeab526e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.041764] env[62460]: DEBUG oslo_vmware.api [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 851.041764] env[62460]: value = "task-1313654" [ 851.041764] env[62460]: _type = "Task" [ 851.041764] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.042917] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 851.042917] env[62460]: value = "task-1313655" [ 851.042917] env[62460]: _type = "Task" [ 851.042917] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.056354] env[62460]: DEBUG oslo_vmware.api [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313654, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.061199] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313655, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.093331] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c2624ccd-5dd5-44c5-9618-3e97c3395a8c tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.873s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.097042] env[62460]: DEBUG oslo_concurrency.lockutils [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.763s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.101021] env[62460]: INFO nova.compute.claims [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 851.125326] env[62460]: INFO nova.scheduler.client.report [None req-c2624ccd-5dd5-44c5-9618-3e97c3395a8c tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Deleted allocations for instance 13890351-6091-4b4f-8484-1cdd0c8523b0 [ 851.141717] env[62460]: DEBUG oslo_vmware.api [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52586062-8e6a-33b1-73e6-6c017285a045, 'name': SearchDatastore_Task, 'duration_secs': 0.015318} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.141717] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-30aea6b1-f18f-49f3-8c2d-99cd20ebf9f0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.145384] env[62460]: DEBUG oslo_vmware.api [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Waiting for the task: (returnval){ [ 851.145384] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52897f28-bc13-6041-0337-266052ecf6d0" [ 851.145384] env[62460]: _type = "Task" [ 851.145384] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.154986] env[62460]: DEBUG oslo_vmware.api [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52897f28-bc13-6041-0337-266052ecf6d0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.223987] env[62460]: DEBUG oslo_vmware.api [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52d698ab-7e57-3dd4-c2a1-8cbb0b07683c, 'name': SearchDatastore_Task, 'duration_secs': 0.030083} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.223987] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.224232] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 851.224447] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.240804] env[62460]: DEBUG nova.compute.manager [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 851.265040] env[62460]: DEBUG nova.virt.hardware [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 851.265301] env[62460]: DEBUG nova.virt.hardware [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 851.265492] env[62460]: DEBUG nova.virt.hardware [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 851.265685] env[62460]: DEBUG nova.virt.hardware [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 851.265860] env[62460]: DEBUG nova.virt.hardware [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 851.266298] env[62460]: DEBUG nova.virt.hardware [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 851.266298] env[62460]: DEBUG nova.virt.hardware [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 851.266411] env[62460]: DEBUG nova.virt.hardware [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 851.266555] env[62460]: DEBUG nova.virt.hardware [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 851.266716] env[62460]: DEBUG nova.virt.hardware [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 851.266946] env[62460]: DEBUG nova.virt.hardware [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 851.268296] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d763cedf-8a26-45c1-a5f3-97da636e5b4b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.274502] env[62460]: DEBUG oslo_concurrency.lockutils [req-7cc1ff3e-848a-4301-ba7a-e0c4a0a11cf1 req-9ff1a9c2-f3b6-45c3-b908-61470d180643 service nova] Releasing lock "refresh_cache-a02b4be1-91b0-4254-8d60-654885e24f6b" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.274890] env[62460]: DEBUG nova.compute.manager [req-7cc1ff3e-848a-4301-ba7a-e0c4a0a11cf1 req-9ff1a9c2-f3b6-45c3-b908-61470d180643 service nova] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Received event network-vif-plugged-bf0f715f-7988-45f9-81b6-c92688ff60c7 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 851.275242] env[62460]: DEBUG oslo_concurrency.lockutils [req-7cc1ff3e-848a-4301-ba7a-e0c4a0a11cf1 req-9ff1a9c2-f3b6-45c3-b908-61470d180643 service nova] Acquiring lock "7e6ff902-4a04-43d5-9014-38c4ec88efc4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.275589] env[62460]: DEBUG oslo_concurrency.lockutils [req-7cc1ff3e-848a-4301-ba7a-e0c4a0a11cf1 req-9ff1a9c2-f3b6-45c3-b908-61470d180643 service nova] Lock "7e6ff902-4a04-43d5-9014-38c4ec88efc4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.275931] env[62460]: DEBUG oslo_concurrency.lockutils [req-7cc1ff3e-848a-4301-ba7a-e0c4a0a11cf1 req-9ff1a9c2-f3b6-45c3-b908-61470d180643 service nova] Lock "7e6ff902-4a04-43d5-9014-38c4ec88efc4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.276133] env[62460]: DEBUG nova.compute.manager [req-7cc1ff3e-848a-4301-ba7a-e0c4a0a11cf1 req-9ff1a9c2-f3b6-45c3-b908-61470d180643 service nova] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] No waiting events found dispatching network-vif-plugged-bf0f715f-7988-45f9-81b6-c92688ff60c7 {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 851.276406] env[62460]: WARNING nova.compute.manager [req-7cc1ff3e-848a-4301-ba7a-e0c4a0a11cf1 req-9ff1a9c2-f3b6-45c3-b908-61470d180643 service nova] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Received unexpected event network-vif-plugged-bf0f715f-7988-45f9-81b6-c92688ff60c7 for instance with vm_state active and task_state None. [ 851.276631] env[62460]: DEBUG nova.compute.manager [req-7cc1ff3e-848a-4301-ba7a-e0c4a0a11cf1 req-9ff1a9c2-f3b6-45c3-b908-61470d180643 service nova] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Received event network-changed-bf0f715f-7988-45f9-81b6-c92688ff60c7 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 851.276852] env[62460]: DEBUG nova.compute.manager [req-7cc1ff3e-848a-4301-ba7a-e0c4a0a11cf1 req-9ff1a9c2-f3b6-45c3-b908-61470d180643 service nova] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Refreshing instance network info cache due to event network-changed-bf0f715f-7988-45f9-81b6-c92688ff60c7. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 851.277098] env[62460]: DEBUG oslo_concurrency.lockutils [req-7cc1ff3e-848a-4301-ba7a-e0c4a0a11cf1 req-9ff1a9c2-f3b6-45c3-b908-61470d180643 service nova] Acquiring lock "refresh_cache-7e6ff902-4a04-43d5-9014-38c4ec88efc4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.277255] env[62460]: DEBUG oslo_concurrency.lockutils [req-7cc1ff3e-848a-4301-ba7a-e0c4a0a11cf1 req-9ff1a9c2-f3b6-45c3-b908-61470d180643 service nova] Acquired lock "refresh_cache-7e6ff902-4a04-43d5-9014-38c4ec88efc4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.277455] env[62460]: DEBUG nova.network.neutron [req-7cc1ff3e-848a-4301-ba7a-e0c4a0a11cf1 req-9ff1a9c2-f3b6-45c3-b908-61470d180643 service nova] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Refreshing network info cache for port bf0f715f-7988-45f9-81b6-c92688ff60c7 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 851.282501] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72eeb231-7a9d-4402-85ed-bc8e5be73aa7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.324045] env[62460]: DEBUG oslo_vmware.api [None req-a781a4a5-2201-450d-aedf-b182703d796e tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313650, 'name': ReconfigVM_Task, 'duration_secs': 0.693469} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.324343] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a781a4a5-2201-450d-aedf-b182703d796e tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Releasing lock "7e6ff902-4a04-43d5-9014-38c4ec88efc4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.324343] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a781a4a5-2201-450d-aedf-b182703d796e tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Reconfigured VM to attach interface {{(pid=62460) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 851.508916] env[62460]: DEBUG oslo_vmware.api [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313651, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07007} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.509260] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 851.510071] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99595af5-1850-43da-82c3-60419e8fd21f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.532058] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] Reconfiguring VM instance instance-00000047 to attach disk [datastore1] 16d40829-ba6e-4193-98b5-fff94c066bcb/16d40829-ba6e-4193-98b5-fff94c066bcb.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 851.532430] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-afc5ea4e-9258-4821-8556-87f2d74a2ef3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.560121] env[62460]: DEBUG oslo_vmware.api [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313654, 'name': CreateSnapshot_Task, 'duration_secs': 0.440847} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.564270] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Created Snapshot of the VM instance {{(pid=62460) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 851.564568] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313655, 'name': CreateVM_Task, 'duration_secs': 0.35168} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.564864] env[62460]: DEBUG oslo_vmware.api [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for the task: (returnval){ [ 851.564864] env[62460]: value = "task-1313656" [ 851.564864] env[62460]: _type = "Task" [ 851.564864] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.565602] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0757eb57-5f54-44df-96e4-df90a2f06e00 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.568105] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 851.568942] env[62460]: DEBUG oslo_concurrency.lockutils [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.569276] env[62460]: DEBUG oslo_concurrency.lockutils [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.569522] env[62460]: DEBUG oslo_concurrency.lockutils [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 851.572560] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b4041a93-ffb3-46b6-a165-dc4708cc5903 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.583048] env[62460]: DEBUG oslo_vmware.api [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313656, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.585106] env[62460]: DEBUG oslo_vmware.api [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Waiting for the task: (returnval){ [ 851.585106] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52ee3de6-6fad-6ec0-c4b2-b924fe44065f" [ 851.585106] env[62460]: _type = "Task" [ 851.585106] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.600346] env[62460]: DEBUG oslo_vmware.api [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52ee3de6-6fad-6ec0-c4b2-b924fe44065f, 'name': SearchDatastore_Task, 'duration_secs': 0.009623} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.600346] env[62460]: DEBUG oslo_concurrency.lockutils [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.600346] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 851.600346] env[62460]: DEBUG oslo_concurrency.lockutils [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.637666] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c2624ccd-5dd5-44c5-9618-3e97c3395a8c tempest-InstanceActionsNegativeTestJSON-1107583005 tempest-InstanceActionsNegativeTestJSON-1107583005-project-member] Lock "13890351-6091-4b4f-8484-1cdd0c8523b0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.551s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.658427] env[62460]: DEBUG oslo_vmware.api [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52897f28-bc13-6041-0337-266052ecf6d0, 'name': SearchDatastore_Task, 'duration_secs': 0.010023} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.658825] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.659220] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] a02b4be1-91b0-4254-8d60-654885e24f6b/a02b4be1-91b0-4254-8d60-654885e24f6b.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 851.659629] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.659921] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 851.660238] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-59123919-7f78-4ff6-939a-e8c1e780f77d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.662575] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ac5e6a79-8461-4942-a982-3a0acd2123e9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.671213] env[62460]: DEBUG oslo_vmware.api [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Waiting for the task: (returnval){ [ 851.671213] env[62460]: value = "task-1313657" [ 851.671213] env[62460]: _type = "Task" [ 851.671213] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.673470] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 851.673764] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 851.674802] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6807c674-0a6c-493b-88ef-0e6ef48f0356 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.680143] env[62460]: DEBUG oslo_vmware.api [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Task: {'id': task-1313657, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.683143] env[62460]: DEBUG oslo_vmware.api [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Waiting for the task: (returnval){ [ 851.683143] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52d6a8c5-8b30-973f-cc7c-059215a06cd5" [ 851.683143] env[62460]: _type = "Task" [ 851.683143] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.694098] env[62460]: DEBUG oslo_vmware.api [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52d6a8c5-8b30-973f-cc7c-059215a06cd5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.829100] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a781a4a5-2201-450d-aedf-b182703d796e tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "interface-7e6ff902-4a04-43d5-9014-38c4ec88efc4-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.821s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.843727] env[62460]: DEBUG nova.network.neutron [req-ad8f3779-3930-475e-83d8-67485ec0063a req-eb8e43bc-d379-4b38-bfe1-3082108484b2 service nova] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] Updated VIF entry in instance network info cache for port 00526c55-f6d3-48d6-ba66-a1f821f2b03b. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 851.844182] env[62460]: DEBUG nova.network.neutron [req-ad8f3779-3930-475e-83d8-67485ec0063a req-eb8e43bc-d379-4b38-bfe1-3082108484b2 service nova] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] Updating instance_info_cache with network_info: [{"id": "00526c55-f6d3-48d6-ba66-a1f821f2b03b", "address": "fa:16:3e:33:a1:71", "network": {"id": "eea7af10-80a3-4d4a-b6d7-0ba4c1780cdf", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-719305021-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "af1a43a6e9824437952b920309572ab3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93341b73-918c-4e9d-9c66-ca171a54b574", "external-id": "nsx-vlan-transportzone-663", "segmentation_id": 663, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap00526c55-f6", "ovs_interfaceid": "00526c55-f6d3-48d6-ba66-a1f821f2b03b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.882038] env[62460]: DEBUG nova.network.neutron [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Successfully updated port: 72a87ddf-0585-429a-b9de-d73bcad42cd1 {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 852.078206] env[62460]: DEBUG oslo_vmware.api [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313656, 'name': ReconfigVM_Task, 'duration_secs': 0.335729} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.078508] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] Reconfigured VM instance instance-00000047 to attach disk [datastore1] 16d40829-ba6e-4193-98b5-fff94c066bcb/16d40829-ba6e-4193-98b5-fff94c066bcb.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 852.079193] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6bdcf156-988f-49bc-aa79-37db32cf756f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.091926] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Creating linked-clone VM from snapshot {{(pid=62460) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 852.093656] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-b982cafe-27db-4799-9b2c-a46ef2e05eb6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.097054] env[62460]: DEBUG oslo_vmware.api [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for the task: (returnval){ [ 852.097054] env[62460]: value = "task-1313658" [ 852.097054] env[62460]: _type = "Task" [ 852.097054] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.105344] env[62460]: DEBUG oslo_vmware.api [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 852.105344] env[62460]: value = "task-1313659" [ 852.105344] env[62460]: _type = "Task" [ 852.105344] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.115489] env[62460]: DEBUG oslo_vmware.api [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313658, 'name': Rename_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.120875] env[62460]: DEBUG oslo_vmware.api [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313659, 'name': CloneVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.178529] env[62460]: DEBUG oslo_vmware.api [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Task: {'id': task-1313657, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.193691] env[62460]: DEBUG oslo_vmware.api [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52d6a8c5-8b30-973f-cc7c-059215a06cd5, 'name': SearchDatastore_Task, 'duration_secs': 0.01705} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.194621] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-60eacf79-e977-4966-8621-53e3b311bde7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.202411] env[62460]: DEBUG oslo_vmware.api [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Waiting for the task: (returnval){ [ 852.202411] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52f12cb1-11c7-71ce-d64b-babaef238e43" [ 852.202411] env[62460]: _type = "Task" [ 852.202411] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.210392] env[62460]: DEBUG oslo_vmware.api [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52f12cb1-11c7-71ce-d64b-babaef238e43, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.339510] env[62460]: DEBUG nova.network.neutron [req-7cc1ff3e-848a-4301-ba7a-e0c4a0a11cf1 req-9ff1a9c2-f3b6-45c3-b908-61470d180643 service nova] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Updated VIF entry in instance network info cache for port bf0f715f-7988-45f9-81b6-c92688ff60c7. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 852.339919] env[62460]: DEBUG nova.network.neutron [req-7cc1ff3e-848a-4301-ba7a-e0c4a0a11cf1 req-9ff1a9c2-f3b6-45c3-b908-61470d180643 service nova] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Updating instance_info_cache with network_info: [{"id": "ba38cbb0-62cf-47cb-a238-87ebc03de250", "address": "fa:16:3e:b4:93:18", "network": {"id": "d3136e32-ad55-4b73-835c-8fa5f0480767", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1511414968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53fa06008e9f43488362895e7a143700", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "054fcd1e-638e-425a-a1de-78cb188ae026", "external-id": "nsx-vlan-transportzone-658", "segmentation_id": 658, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba38cbb0-62", "ovs_interfaceid": "ba38cbb0-62cf-47cb-a238-87ebc03de250", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "bf0f715f-7988-45f9-81b6-c92688ff60c7", "address": "fa:16:3e:84:3c:55", "network": {"id": "d3136e32-ad55-4b73-835c-8fa5f0480767", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1511414968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53fa06008e9f43488362895e7a143700", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "054fcd1e-638e-425a-a1de-78cb188ae026", "external-id": "nsx-vlan-transportzone-658", "segmentation_id": 658, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf0f715f-79", "ovs_interfaceid": "bf0f715f-7988-45f9-81b6-c92688ff60c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.349834] env[62460]: DEBUG oslo_concurrency.lockutils [req-ad8f3779-3930-475e-83d8-67485ec0063a req-eb8e43bc-d379-4b38-bfe1-3082108484b2 service nova] Releasing lock "refresh_cache-49477d35-92ea-4f9d-8333-fc23144b7dfc" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.385314] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquiring lock "refresh_cache-896f8ff4-e45f-4403-a727-03ee25e58609" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.385447] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquired lock "refresh_cache-896f8ff4-e45f-4403-a727-03ee25e58609" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.386080] env[62460]: DEBUG nova.network.neutron [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 852.407191] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eba2a63b-aee6-4ee0-92df-54f92dcd24a1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.415970] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db2ba27a-eb22-4d0d-8df9-f98a1a87e868 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.448319] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5db477a3-ef78-4b1e-abdc-68157652a7d0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.456247] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-030d97a0-275e-4bee-915c-d20b9daaba84 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.470285] env[62460]: DEBUG nova.compute.provider_tree [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 852.597141] env[62460]: DEBUG nova.compute.manager [req-806d3a0c-b2e7-4708-8604-37e24fca4b5f req-d6ef5500-bf6b-46d4-8012-29b8e491e954 service nova] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Received event network-vif-plugged-72a87ddf-0585-429a-b9de-d73bcad42cd1 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 852.597468] env[62460]: DEBUG oslo_concurrency.lockutils [req-806d3a0c-b2e7-4708-8604-37e24fca4b5f req-d6ef5500-bf6b-46d4-8012-29b8e491e954 service nova] Acquiring lock "896f8ff4-e45f-4403-a727-03ee25e58609-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.597573] env[62460]: DEBUG oslo_concurrency.lockutils [req-806d3a0c-b2e7-4708-8604-37e24fca4b5f req-d6ef5500-bf6b-46d4-8012-29b8e491e954 service nova] Lock "896f8ff4-e45f-4403-a727-03ee25e58609-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.597815] env[62460]: DEBUG oslo_concurrency.lockutils [req-806d3a0c-b2e7-4708-8604-37e24fca4b5f req-d6ef5500-bf6b-46d4-8012-29b8e491e954 service nova] Lock "896f8ff4-e45f-4403-a727-03ee25e58609-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.599727] env[62460]: DEBUG nova.compute.manager [req-806d3a0c-b2e7-4708-8604-37e24fca4b5f req-d6ef5500-bf6b-46d4-8012-29b8e491e954 service nova] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] No waiting events found dispatching network-vif-plugged-72a87ddf-0585-429a-b9de-d73bcad42cd1 {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 852.599993] env[62460]: WARNING nova.compute.manager [req-806d3a0c-b2e7-4708-8604-37e24fca4b5f req-d6ef5500-bf6b-46d4-8012-29b8e491e954 service nova] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Received unexpected event network-vif-plugged-72a87ddf-0585-429a-b9de-d73bcad42cd1 for instance with vm_state building and task_state spawning. [ 852.600154] env[62460]: DEBUG nova.compute.manager [req-806d3a0c-b2e7-4708-8604-37e24fca4b5f req-d6ef5500-bf6b-46d4-8012-29b8e491e954 service nova] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Received event network-changed-72a87ddf-0585-429a-b9de-d73bcad42cd1 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 852.600444] env[62460]: DEBUG nova.compute.manager [req-806d3a0c-b2e7-4708-8604-37e24fca4b5f req-d6ef5500-bf6b-46d4-8012-29b8e491e954 service nova] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Refreshing instance network info cache due to event network-changed-72a87ddf-0585-429a-b9de-d73bcad42cd1. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 852.600805] env[62460]: DEBUG oslo_concurrency.lockutils [req-806d3a0c-b2e7-4708-8604-37e24fca4b5f req-d6ef5500-bf6b-46d4-8012-29b8e491e954 service nova] Acquiring lock "refresh_cache-896f8ff4-e45f-4403-a727-03ee25e58609" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.615319] env[62460]: DEBUG oslo_vmware.api [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313658, 'name': Rename_Task, 'duration_secs': 0.439817} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.616311] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 852.616617] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e8cf17fa-741c-4ac8-bb31-fc248188dbe2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.622232] env[62460]: DEBUG oslo_vmware.api [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313659, 'name': CloneVM_Task} progress is 94%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.628423] env[62460]: DEBUG oslo_vmware.api [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for the task: (returnval){ [ 852.628423] env[62460]: value = "task-1313660" [ 852.628423] env[62460]: _type = "Task" [ 852.628423] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.637666] env[62460]: DEBUG oslo_vmware.api [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313660, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.685530] env[62460]: DEBUG oslo_vmware.api [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Task: {'id': task-1313657, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.525773} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.686388] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] a02b4be1-91b0-4254-8d60-654885e24f6b/a02b4be1-91b0-4254-8d60-654885e24f6b.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 852.687042] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 852.687662] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-75138d3c-a2fc-4c6c-95ad-21823ba2eb59 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.697030] env[62460]: DEBUG oslo_vmware.api [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Waiting for the task: (returnval){ [ 852.697030] env[62460]: value = "task-1313661" [ 852.697030] env[62460]: _type = "Task" [ 852.697030] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.717130] env[62460]: DEBUG oslo_vmware.api [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Task: {'id': task-1313661, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.724777] env[62460]: DEBUG oslo_vmware.api [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52f12cb1-11c7-71ce-d64b-babaef238e43, 'name': SearchDatastore_Task, 'duration_secs': 0.016049} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.724777] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.725402] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 9d1a5830-f3c0-4d18-9338-16f7b6962c6a/9d1a5830-f3c0-4d18-9338-16f7b6962c6a.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 852.725942] env[62460]: DEBUG oslo_concurrency.lockutils [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.726360] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 852.726769] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-79168daa-2af7-453f-ac70-a2db19cc30a3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.730200] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0d7a6117-5d5f-43ef-9a6d-17c721eb062c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.742862] env[62460]: DEBUG oslo_vmware.api [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Waiting for the task: (returnval){ [ 852.742862] env[62460]: value = "task-1313662" [ 852.742862] env[62460]: _type = "Task" [ 852.742862] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.750428] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 852.750809] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 852.752447] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2a2ebd16-72bb-4fc5-9a7b-8aa2c4f37568 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.759114] env[62460]: DEBUG oslo_vmware.api [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Task: {'id': task-1313662, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.762406] env[62460]: DEBUG oslo_vmware.api [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Waiting for the task: (returnval){ [ 852.762406] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52cc2a7f-230f-88c3-2ba3-113e9886ef5e" [ 852.762406] env[62460]: _type = "Task" [ 852.762406] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.770548] env[62460]: DEBUG oslo_vmware.api [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52cc2a7f-230f-88c3-2ba3-113e9886ef5e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.842448] env[62460]: DEBUG oslo_concurrency.lockutils [req-7cc1ff3e-848a-4301-ba7a-e0c4a0a11cf1 req-9ff1a9c2-f3b6-45c3-b908-61470d180643 service nova] Releasing lock "refresh_cache-7e6ff902-4a04-43d5-9014-38c4ec88efc4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.844195] env[62460]: DEBUG nova.compute.manager [req-7cc1ff3e-848a-4301-ba7a-e0c4a0a11cf1 req-9ff1a9c2-f3b6-45c3-b908-61470d180643 service nova] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] Received event network-vif-deleted-608a7ef6-983b-4023-a6b6-69e632436554 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 852.935611] env[62460]: DEBUG nova.network.neutron [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 852.994679] env[62460]: ERROR nova.scheduler.client.report [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] [req-a6291692-3d8e-4fe8-b297-f313e698bd0a] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 76e8f00d-65cc-4766-bb9d-5006fb51c728. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-a6291692-3d8e-4fe8-b297-f313e698bd0a"}]} [ 853.027964] env[62460]: DEBUG nova.scheduler.client.report [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Refreshing inventories for resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 853.053313] env[62460]: DEBUG nova.scheduler.client.report [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Updating ProviderTree inventory for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 853.054138] env[62460]: DEBUG nova.compute.provider_tree [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 853.071572] env[62460]: DEBUG nova.scheduler.client.report [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Refreshing aggregate associations for resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728, aggregates: None {{(pid=62460) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 853.102517] env[62460]: DEBUG nova.scheduler.client.report [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Refreshing trait associations for resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62460) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 853.117179] env[62460]: DEBUG oslo_vmware.api [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313659, 'name': CloneVM_Task} progress is 94%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.147501] env[62460]: DEBUG oslo_vmware.api [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313660, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.210289] env[62460]: DEBUG oslo_vmware.api [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Task: {'id': task-1313661, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.170313} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.210524] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 853.211421] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61739107-965c-4a16-9cc6-0d688c3d36eb {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.245275] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Reconfiguring VM instance instance-00000048 to attach disk [datastore1] a02b4be1-91b0-4254-8d60-654885e24f6b/a02b4be1-91b0-4254-8d60-654885e24f6b.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 853.250314] env[62460]: DEBUG nova.network.neutron [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Updating instance_info_cache with network_info: [{"id": "72a87ddf-0585-429a-b9de-d73bcad42cd1", "address": "fa:16:3e:85:2f:50", "network": {"id": "b5cb4ee0-4562-45be-a25d-51c3a5932efd", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-823340443-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3d2d1c48ec14121a2e8c9b3f800a949", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55c757ac-f8b2-466d-b634-07dbd100b312", "external-id": "nsx-vlan-transportzone-159", "segmentation_id": 159, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap72a87ddf-05", "ovs_interfaceid": "72a87ddf-0585-429a-b9de-d73bcad42cd1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 853.252711] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ff8a33ac-d102-4b39-93dc-0071fca936be {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.287046] env[62460]: DEBUG oslo_vmware.api [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Task: {'id': task-1313662, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.294856] env[62460]: DEBUG oslo_vmware.api [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Waiting for the task: (returnval){ [ 853.294856] env[62460]: value = "task-1313663" [ 853.294856] env[62460]: _type = "Task" [ 853.294856] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.295759] env[62460]: DEBUG oslo_vmware.api [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52cc2a7f-230f-88c3-2ba3-113e9886ef5e, 'name': SearchDatastore_Task, 'duration_secs': 0.030868} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.300108] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ebe3744-c665-4860-a012-1b0e45d94002 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.308808] env[62460]: DEBUG oslo_vmware.api [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Task: {'id': task-1313663, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.313721] env[62460]: DEBUG oslo_vmware.api [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Waiting for the task: (returnval){ [ 853.313721] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]5202af49-050e-db63-3910-0aaf43d84c92" [ 853.313721] env[62460]: _type = "Task" [ 853.313721] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.325431] env[62460]: DEBUG oslo_vmware.api [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5202af49-050e-db63-3910-0aaf43d84c92, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.509262] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ac6772a-21da-43ff-8f99-708a7baa1194 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.518454] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30cd5c33-6f62-4964-b4ff-aba2117b8301 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.555093] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7498c5f3-9db2-4080-b6d4-dec747522399 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.569744] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a959d14-59f4-442e-93ee-d93309c34d9b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.596310] env[62460]: DEBUG nova.compute.provider_tree [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 853.619665] env[62460]: DEBUG oslo_vmware.api [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313659, 'name': CloneVM_Task} progress is 94%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.643590] env[62460]: DEBUG oslo_vmware.api [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313660, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.754349] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Releasing lock "refresh_cache-896f8ff4-e45f-4403-a727-03ee25e58609" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.754349] env[62460]: DEBUG nova.compute.manager [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Instance network_info: |[{"id": "72a87ddf-0585-429a-b9de-d73bcad42cd1", "address": "fa:16:3e:85:2f:50", "network": {"id": "b5cb4ee0-4562-45be-a25d-51c3a5932efd", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-823340443-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3d2d1c48ec14121a2e8c9b3f800a949", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55c757ac-f8b2-466d-b634-07dbd100b312", "external-id": "nsx-vlan-transportzone-159", "segmentation_id": 159, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap72a87ddf-05", "ovs_interfaceid": "72a87ddf-0585-429a-b9de-d73bcad42cd1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 853.755120] env[62460]: DEBUG oslo_concurrency.lockutils [req-806d3a0c-b2e7-4708-8604-37e24fca4b5f req-d6ef5500-bf6b-46d4-8012-29b8e491e954 service nova] Acquired lock "refresh_cache-896f8ff4-e45f-4403-a727-03ee25e58609" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.755273] env[62460]: DEBUG nova.network.neutron [req-806d3a0c-b2e7-4708-8604-37e24fca4b5f req-d6ef5500-bf6b-46d4-8012-29b8e491e954 service nova] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Refreshing network info cache for port 72a87ddf-0585-429a-b9de-d73bcad42cd1 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 853.757120] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:85:2f:50', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '55c757ac-f8b2-466d-b634-07dbd100b312', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '72a87ddf-0585-429a-b9de-d73bcad42cd1', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 853.767057] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Creating folder: Project (f3d2d1c48ec14121a2e8c9b3f800a949). Parent ref: group-v281134. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 853.772099] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-36d75314-74f2-4214-9313-7ca6a409f03c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.781275] env[62460]: DEBUG oslo_vmware.api [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Task: {'id': task-1313662, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.890033} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.781648] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 9d1a5830-f3c0-4d18-9338-16f7b6962c6a/9d1a5830-f3c0-4d18-9338-16f7b6962c6a.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 853.781918] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 853.782229] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-64765363-1a4a-466f-998f-519cfc9c3ceb {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.787268] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Created folder: Project (f3d2d1c48ec14121a2e8c9b3f800a949) in parent group-v281134. [ 853.787268] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Creating folder: Instances. Parent ref: group-v281209. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 853.787987] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8f67fd6c-e8d4-4fbb-be1f-9a0f5b65f42f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.792097] env[62460]: DEBUG oslo_vmware.api [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Waiting for the task: (returnval){ [ 853.792097] env[62460]: value = "task-1313665" [ 853.792097] env[62460]: _type = "Task" [ 853.792097] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.805671] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Created folder: Instances in parent group-v281209. [ 853.806212] env[62460]: DEBUG oslo.service.loopingcall [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 853.813436] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 853.813920] env[62460]: DEBUG oslo_vmware.api [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Task: {'id': task-1313665, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.814308] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6ecb14c7-60d5-44ae-96cb-b5a5a55647fe {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.841185] env[62460]: DEBUG oslo_vmware.api [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Task: {'id': task-1313663, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.853032] env[62460]: DEBUG oslo_vmware.api [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5202af49-050e-db63-3910-0aaf43d84c92, 'name': SearchDatastore_Task, 'duration_secs': 0.054029} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.853194] env[62460]: DEBUG oslo_concurrency.lockutils [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.856489] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 49477d35-92ea-4f9d-8333-fc23144b7dfc/49477d35-92ea-4f9d-8333-fc23144b7dfc.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 853.856489] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 853.856489] env[62460]: value = "task-1313667" [ 853.856489] env[62460]: _type = "Task" [ 853.856489] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.856489] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4123f3b4-c3ee-4d25-807b-1c9b063df680 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.868218] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313667, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.870394] env[62460]: DEBUG oslo_vmware.api [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Waiting for the task: (returnval){ [ 853.870394] env[62460]: value = "task-1313668" [ 853.870394] env[62460]: _type = "Task" [ 853.870394] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.882963] env[62460]: DEBUG oslo_vmware.api [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Task: {'id': task-1313668, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.062394] env[62460]: DEBUG oslo_concurrency.lockutils [None req-44d737cd-9ee5-48a0-baac-6819166b7570 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "interface-7e6ff902-4a04-43d5-9014-38c4ec88efc4-d1397bb8-e534-405e-9921-e3e7e6914f81" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.063238] env[62460]: DEBUG oslo_concurrency.lockutils [None req-44d737cd-9ee5-48a0-baac-6819166b7570 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "interface-7e6ff902-4a04-43d5-9014-38c4ec88efc4-d1397bb8-e534-405e-9921-e3e7e6914f81" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.063238] env[62460]: DEBUG nova.objects.instance [None req-44d737cd-9ee5-48a0-baac-6819166b7570 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lazy-loading 'flavor' on Instance uuid 7e6ff902-4a04-43d5-9014-38c4ec88efc4 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 854.101592] env[62460]: DEBUG nova.scheduler.client.report [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 854.122248] env[62460]: DEBUG oslo_vmware.api [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313659, 'name': CloneVM_Task, 'duration_secs': 1.949503} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.122637] env[62460]: INFO nova.virt.vmwareapi.vmops [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Created linked-clone VM from snapshot [ 854.123454] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-457fa601-85f2-474e-ad95-dc737567bdc9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.138081] env[62460]: DEBUG nova.virt.vmwareapi.images [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Uploading image 85c60cb8-8b28-497d-afb9-ec82145a4f6e {{(pid=62460) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 854.152615] env[62460]: DEBUG oslo_vmware.api [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313660, 'name': PowerOnVM_Task, 'duration_secs': 1.162591} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.153467] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 854.153467] env[62460]: INFO nova.compute.manager [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] Took 10.07 seconds to spawn the instance on the hypervisor. [ 854.153467] env[62460]: DEBUG nova.compute.manager [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 854.154352] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-beb61d8e-aea9-4af2-92e1-131fdc77f38d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.178991] env[62460]: DEBUG oslo_vmware.rw_handles [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 854.178991] env[62460]: value = "vm-281208" [ 854.178991] env[62460]: _type = "VirtualMachine" [ 854.178991] env[62460]: }. {{(pid=62460) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 854.179453] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-0c8a2aae-b256-40c2-ad37-fb9196367480 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.189285] env[62460]: DEBUG oslo_vmware.rw_handles [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lease: (returnval){ [ 854.189285] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]522da242-84ec-b198-3053-1058e6e17018" [ 854.189285] env[62460]: _type = "HttpNfcLease" [ 854.189285] env[62460]: } obtained for exporting VM: (result){ [ 854.189285] env[62460]: value = "vm-281208" [ 854.189285] env[62460]: _type = "VirtualMachine" [ 854.189285] env[62460]: }. {{(pid=62460) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 854.189285] env[62460]: DEBUG oslo_vmware.api [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the lease: (returnval){ [ 854.189285] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]522da242-84ec-b198-3053-1058e6e17018" [ 854.189285] env[62460]: _type = "HttpNfcLease" [ 854.189285] env[62460]: } to be ready. {{(pid=62460) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 854.197593] env[62460]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 854.197593] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]522da242-84ec-b198-3053-1058e6e17018" [ 854.197593] env[62460]: _type = "HttpNfcLease" [ 854.197593] env[62460]: } is initializing. {{(pid=62460) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 854.305426] env[62460]: DEBUG oslo_vmware.api [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Task: {'id': task-1313665, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080284} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.306548] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 854.307700] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-420bd825-20e9-4c83-b1b6-d393fdb62358 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.314185] env[62460]: DEBUG oslo_vmware.api [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Task: {'id': task-1313663, 'name': ReconfigVM_Task, 'duration_secs': 0.640534} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.314938] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Reconfigured VM instance instance-00000048 to attach disk [datastore1] a02b4be1-91b0-4254-8d60-654885e24f6b/a02b4be1-91b0-4254-8d60-654885e24f6b.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 854.315805] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-88370232-2952-4635-b6b4-56c617541c71 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.334708] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Reconfiguring VM instance instance-00000045 to attach disk [datastore1] 9d1a5830-f3c0-4d18-9338-16f7b6962c6a/9d1a5830-f3c0-4d18-9338-16f7b6962c6a.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 854.335597] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7adf882a-fb3a-453b-8c30-1640b098103a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.355952] env[62460]: DEBUG oslo_vmware.api [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Waiting for the task: (returnval){ [ 854.355952] env[62460]: value = "task-1313670" [ 854.355952] env[62460]: _type = "Task" [ 854.355952] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.362852] env[62460]: DEBUG oslo_vmware.api [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Waiting for the task: (returnval){ [ 854.362852] env[62460]: value = "task-1313671" [ 854.362852] env[62460]: _type = "Task" [ 854.362852] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.373854] env[62460]: DEBUG oslo_vmware.api [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Task: {'id': task-1313670, 'name': Rename_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.387525] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313667, 'name': CreateVM_Task, 'duration_secs': 0.508173} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.387525] env[62460]: DEBUG oslo_vmware.api [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Task: {'id': task-1313671, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.388114] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 854.388865] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.388931] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.389332] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 854.392811] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-99854477-1175-440e-b68b-e22526589f0c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.396569] env[62460]: DEBUG oslo_vmware.api [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Task: {'id': task-1313668, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.398216] env[62460]: DEBUG oslo_vmware.api [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 854.398216] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]522aa6cb-58e1-82d4-844a-784110f3a86f" [ 854.398216] env[62460]: _type = "Task" [ 854.398216] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.406539] env[62460]: DEBUG oslo_vmware.api [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]522aa6cb-58e1-82d4-844a-784110f3a86f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.552570] env[62460]: DEBUG nova.network.neutron [req-806d3a0c-b2e7-4708-8604-37e24fca4b5f req-d6ef5500-bf6b-46d4-8012-29b8e491e954 service nova] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Updated VIF entry in instance network info cache for port 72a87ddf-0585-429a-b9de-d73bcad42cd1. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 854.552764] env[62460]: DEBUG nova.network.neutron [req-806d3a0c-b2e7-4708-8604-37e24fca4b5f req-d6ef5500-bf6b-46d4-8012-29b8e491e954 service nova] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Updating instance_info_cache with network_info: [{"id": "72a87ddf-0585-429a-b9de-d73bcad42cd1", "address": "fa:16:3e:85:2f:50", "network": {"id": "b5cb4ee0-4562-45be-a25d-51c3a5932efd", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-823340443-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3d2d1c48ec14121a2e8c9b3f800a949", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55c757ac-f8b2-466d-b634-07dbd100b312", "external-id": "nsx-vlan-transportzone-159", "segmentation_id": 159, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap72a87ddf-05", "ovs_interfaceid": "72a87ddf-0585-429a-b9de-d73bcad42cd1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.610255] env[62460]: DEBUG oslo_concurrency.lockutils [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.514s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.610809] env[62460]: DEBUG nova.compute.manager [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 854.614323] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.720s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.615718] env[62460]: INFO nova.compute.claims [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 854.684070] env[62460]: INFO nova.compute.manager [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] Took 32.57 seconds to build instance. [ 854.698382] env[62460]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 854.698382] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]522da242-84ec-b198-3053-1058e6e17018" [ 854.698382] env[62460]: _type = "HttpNfcLease" [ 854.698382] env[62460]: } is ready. {{(pid=62460) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 854.699152] env[62460]: DEBUG nova.objects.instance [None req-44d737cd-9ee5-48a0-baac-6819166b7570 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lazy-loading 'pci_requests' on Instance uuid 7e6ff902-4a04-43d5-9014-38c4ec88efc4 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 854.700613] env[62460]: DEBUG oslo_vmware.rw_handles [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 854.700613] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]522da242-84ec-b198-3053-1058e6e17018" [ 854.700613] env[62460]: _type = "HttpNfcLease" [ 854.700613] env[62460]: }. {{(pid=62460) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 854.701577] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-336366cc-1120-49de-810f-50feec0979b4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.712611] env[62460]: DEBUG oslo_vmware.rw_handles [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52542c93-0f0c-d878-684e-d5939a14b76c/disk-0.vmdk from lease info. {{(pid=62460) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 854.712800] env[62460]: DEBUG oslo_vmware.rw_handles [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52542c93-0f0c-d878-684e-d5939a14b76c/disk-0.vmdk for reading. {{(pid=62460) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 854.863766] env[62460]: DEBUG oslo_vmware.api [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Task: {'id': task-1313670, 'name': Rename_Task, 'duration_secs': 0.485476} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.864114] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 854.864410] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7ad94824-cc2b-478d-b65a-207abaaa48fc {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.872299] env[62460]: DEBUG oslo_vmware.api [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Waiting for the task: (returnval){ [ 854.872299] env[62460]: value = "task-1313672" [ 854.872299] env[62460]: _type = "Task" [ 854.872299] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.884499] env[62460]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-c30bd26a-ffa8-48ac-85f7-a263e29102ab {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.886221] env[62460]: DEBUG oslo_vmware.api [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Task: {'id': task-1313671, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.889697] env[62460]: DEBUG oslo_vmware.api [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Task: {'id': task-1313672, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.894518] env[62460]: DEBUG oslo_vmware.api [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Task: {'id': task-1313668, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.606402} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.897337] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 49477d35-92ea-4f9d-8333-fc23144b7dfc/49477d35-92ea-4f9d-8333-fc23144b7dfc.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 854.897548] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 854.898390] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-59339124-9a9f-44e0-8c47-524346e80756 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.908783] env[62460]: DEBUG oslo_vmware.api [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Waiting for the task: (returnval){ [ 854.908783] env[62460]: value = "task-1313673" [ 854.908783] env[62460]: _type = "Task" [ 854.908783] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.915940] env[62460]: DEBUG oslo_vmware.api [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]522aa6cb-58e1-82d4-844a-784110f3a86f, 'name': SearchDatastore_Task, 'duration_secs': 0.044809} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.922730] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.923081] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 854.923274] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.923461] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.923623] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 854.927144] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cfaf6c20-187f-46bf-89bd-327cb4ac73ab {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.929094] env[62460]: DEBUG oslo_vmware.api [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Task: {'id': task-1313673, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.935995] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 854.936232] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 854.936999] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9f3742fc-607f-4f57-9163-777cc3a4e481 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.941923] env[62460]: DEBUG oslo_vmware.api [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 854.941923] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]5278b4ee-0391-23fe-ae81-da29f9f22a83" [ 854.941923] env[62460]: _type = "Task" [ 854.941923] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.950665] env[62460]: DEBUG oslo_vmware.api [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5278b4ee-0391-23fe-ae81-da29f9f22a83, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.055376] env[62460]: DEBUG oslo_concurrency.lockutils [req-806d3a0c-b2e7-4708-8604-37e24fca4b5f req-d6ef5500-bf6b-46d4-8012-29b8e491e954 service nova] Releasing lock "refresh_cache-896f8ff4-e45f-4403-a727-03ee25e58609" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.120154] env[62460]: DEBUG nova.compute.utils [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 855.139379] env[62460]: DEBUG nova.compute.manager [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 855.141430] env[62460]: DEBUG nova.network.neutron [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 855.151303] env[62460]: DEBUG nova.compute.manager [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 855.186557] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ad4680a7-74f0-4d93-b3fa-2d136b1bb6ab tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "16d40829-ba6e-4193-98b5-fff94c066bcb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 78.538s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.202530] env[62460]: DEBUG nova.objects.base [None req-44d737cd-9ee5-48a0-baac-6819166b7570 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Object Instance<7e6ff902-4a04-43d5-9014-38c4ec88efc4> lazy-loaded attributes: flavor,pci_requests {{(pid=62460) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 855.203010] env[62460]: DEBUG nova.network.neutron [None req-44d737cd-9ee5-48a0-baac-6819166b7570 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 855.209514] env[62460]: DEBUG nova.policy [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '363d5f1a4f394c8c91802a97cdea237a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a10038b038254e5d9543018e39a56992', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 855.299281] env[62460]: DEBUG nova.policy [None req-44d737cd-9ee5-48a0-baac-6819166b7570 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd7b38fb0a53f4d078ae86d68e3eb6268', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '53fa06008e9f43488362895e7a143700', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 855.380694] env[62460]: DEBUG oslo_vmware.api [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Task: {'id': task-1313671, 'name': ReconfigVM_Task, 'duration_secs': 0.527559} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.381953] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Reconfigured VM instance instance-00000045 to attach disk [datastore1] 9d1a5830-f3c0-4d18-9338-16f7b6962c6a/9d1a5830-f3c0-4d18-9338-16f7b6962c6a.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 855.382760] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c3c99503-387d-40b5-b540-68c546ae2d2f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.388355] env[62460]: DEBUG oslo_vmware.api [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Task: {'id': task-1313672, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.394593] env[62460]: DEBUG oslo_vmware.api [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Waiting for the task: (returnval){ [ 855.394593] env[62460]: value = "task-1313674" [ 855.394593] env[62460]: _type = "Task" [ 855.394593] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.408028] env[62460]: DEBUG oslo_vmware.api [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Task: {'id': task-1313674, 'name': Rename_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.417898] env[62460]: DEBUG oslo_vmware.api [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Task: {'id': task-1313673, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067785} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.418345] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 855.419322] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2466692a-f57c-44cb-994d-4837b844e4e4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.446600] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] Reconfiguring VM instance instance-00000049 to attach disk [datastore1] 49477d35-92ea-4f9d-8333-fc23144b7dfc/49477d35-92ea-4f9d-8333-fc23144b7dfc.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 855.447060] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-02661633-1bad-47dc-9126-106c18f50d4d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.474284] env[62460]: DEBUG oslo_vmware.api [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5278b4ee-0391-23fe-ae81-da29f9f22a83, 'name': SearchDatastore_Task, 'duration_secs': 0.024421} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.477286] env[62460]: DEBUG oslo_vmware.api [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Waiting for the task: (returnval){ [ 855.477286] env[62460]: value = "task-1313675" [ 855.477286] env[62460]: _type = "Task" [ 855.477286] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.477286] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e1ceef89-373c-4e4b-ace4-fdf1bc8c2891 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.487152] env[62460]: DEBUG oslo_vmware.api [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 855.487152] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52412db9-bcc8-b46b-4dd4-bfe856d3eaa7" [ 855.487152] env[62460]: _type = "Task" [ 855.487152] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.489108] env[62460]: DEBUG oslo_vmware.api [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Task: {'id': task-1313675, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.498430] env[62460]: DEBUG oslo_vmware.api [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52412db9-bcc8-b46b-4dd4-bfe856d3eaa7, 'name': SearchDatastore_Task, 'duration_secs': 0.010601} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.498735] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.498997] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 896f8ff4-e45f-4403-a727-03ee25e58609/896f8ff4-e45f-4403-a727-03ee25e58609.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 855.499522] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-21baab0f-513e-4a30-b5e3-b32fe652f21b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.505170] env[62460]: DEBUG oslo_vmware.api [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 855.505170] env[62460]: value = "task-1313676" [ 855.505170] env[62460]: _type = "Task" [ 855.505170] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.513573] env[62460]: DEBUG oslo_vmware.api [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313676, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.534935] env[62460]: DEBUG nova.network.neutron [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Successfully created port: dbacff36-64b1-4fc7-b0a0-dc78fcb72ac2 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 855.656344] env[62460]: INFO nova.virt.block_device [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Booting with volume 2cf093c7-9278-4f2d-a385-8775acc7dae8 at /dev/sda [ 855.702701] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bb633480-f536-4aa4-b07f-c5329f887c19 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.712045] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1ec3cb8-25fb-4140-a6e7-34931ed72011 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.749840] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c00d2686-eb21-4f78-b62b-2d69edc8a7b6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.760240] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33293276-5427-4648-897e-50b07d56ed1b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.802555] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b4a5a88-6d61-40ae-bcab-5f0b0f1e136a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.812486] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d45fbd5-6324-44b3-8cf3-486bc6a9f3bf {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.829265] env[62460]: DEBUG nova.virt.block_device [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Updating existing volume attachment record: 857b0225-8e87-4ba9-b6c1-d9ab66b4754b {{(pid=62460) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 855.885368] env[62460]: DEBUG oslo_vmware.api [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Task: {'id': task-1313672, 'name': PowerOnVM_Task, 'duration_secs': 0.798892} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.888437] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 855.888667] env[62460]: INFO nova.compute.manager [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Took 9.45 seconds to spawn the instance on the hypervisor. [ 855.888867] env[62460]: DEBUG nova.compute.manager [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 855.890168] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8052fb58-f72c-438a-8aaf-6eaa772dcd49 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.909535] env[62460]: DEBUG oslo_vmware.api [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Task: {'id': task-1313674, 'name': Rename_Task, 'duration_secs': 0.139263} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.909535] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 855.909535] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ec651ae3-048e-4b48-90e3-dc15d2b6c0c8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.918170] env[62460]: DEBUG oslo_vmware.api [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Waiting for the task: (returnval){ [ 855.918170] env[62460]: value = "task-1313677" [ 855.918170] env[62460]: _type = "Task" [ 855.918170] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.927694] env[62460]: DEBUG oslo_vmware.api [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Task: {'id': task-1313677, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.992893] env[62460]: DEBUG oslo_vmware.api [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Task: {'id': task-1313675, 'name': ReconfigVM_Task, 'duration_secs': 0.276578} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.993216] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] Reconfigured VM instance instance-00000049 to attach disk [datastore1] 49477d35-92ea-4f9d-8333-fc23144b7dfc/49477d35-92ea-4f9d-8333-fc23144b7dfc.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 855.999909] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c39ef814-cb12-46b8-a61e-788b5b30fcd3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.008974] env[62460]: DEBUG oslo_vmware.api [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Waiting for the task: (returnval){ [ 856.008974] env[62460]: value = "task-1313678" [ 856.008974] env[62460]: _type = "Task" [ 856.008974] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.024769] env[62460]: DEBUG oslo_vmware.api [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313676, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.028728] env[62460]: DEBUG oslo_vmware.api [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Task: {'id': task-1313678, 'name': Rename_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.052626] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e178253d-5ccb-4ba9-b122-324a25df1bdf {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.062709] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3106b9d7-9de8-43f2-8051-899b3cf76bc5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.095617] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a2f60fe-fce3-4a9e-9e95-614988b5dd4f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.104135] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10285976-8498-419a-9567-d77220b09e71 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.118606] env[62460]: DEBUG nova.compute.provider_tree [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 856.412923] env[62460]: INFO nova.compute.manager [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Took 31.98 seconds to build instance. [ 856.430975] env[62460]: DEBUG oslo_vmware.api [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Task: {'id': task-1313677, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.450497] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5770ec75-799a-4db6-97a5-b98712f6f11d tempest-ServersListShow296Test-180931441 tempest-ServersListShow296Test-180931441-project-member] Acquiring lock "120b7508-dc38-4659-92cb-8b3e813dae2a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.450774] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5770ec75-799a-4db6-97a5-b98712f6f11d tempest-ServersListShow296Test-180931441 tempest-ServersListShow296Test-180931441-project-member] Lock "120b7508-dc38-4659-92cb-8b3e813dae2a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.480499] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a0a3bf88-42e5-4bc9-9abb-b76cbcd709f3 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "16d40829-ba6e-4193-98b5-fff94c066bcb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.480714] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a0a3bf88-42e5-4bc9-9abb-b76cbcd709f3 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "16d40829-ba6e-4193-98b5-fff94c066bcb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.480945] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a0a3bf88-42e5-4bc9-9abb-b76cbcd709f3 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "16d40829-ba6e-4193-98b5-fff94c066bcb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.481231] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a0a3bf88-42e5-4bc9-9abb-b76cbcd709f3 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "16d40829-ba6e-4193-98b5-fff94c066bcb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.481417] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a0a3bf88-42e5-4bc9-9abb-b76cbcd709f3 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "16d40829-ba6e-4193-98b5-fff94c066bcb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.483983] env[62460]: INFO nova.compute.manager [None req-a0a3bf88-42e5-4bc9-9abb-b76cbcd709f3 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] Terminating instance [ 856.485892] env[62460]: DEBUG nova.compute.manager [None req-a0a3bf88-42e5-4bc9-9abb-b76cbcd709f3 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 856.486204] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a0a3bf88-42e5-4bc9-9abb-b76cbcd709f3 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 856.487053] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41b1a9c7-1dd4-4643-854c-5123c330edec {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.494898] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0a3bf88-42e5-4bc9-9abb-b76cbcd709f3 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 856.495218] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-492c580a-501d-42c7-9d4f-aa3294fff8a9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.501620] env[62460]: DEBUG oslo_vmware.api [None req-a0a3bf88-42e5-4bc9-9abb-b76cbcd709f3 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for the task: (returnval){ [ 856.501620] env[62460]: value = "task-1313679" [ 856.501620] env[62460]: _type = "Task" [ 856.501620] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.509583] env[62460]: DEBUG oslo_vmware.api [None req-a0a3bf88-42e5-4bc9-9abb-b76cbcd709f3 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313679, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.520193] env[62460]: DEBUG oslo_vmware.api [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313676, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.585823} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.520837] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 896f8ff4-e45f-4403-a727-03ee25e58609/896f8ff4-e45f-4403-a727-03ee25e58609.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 856.521079] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 856.521355] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-900f12a9-73a8-44b5-b8bf-7f47405c6489 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.527560] env[62460]: DEBUG oslo_vmware.api [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Task: {'id': task-1313678, 'name': Rename_Task, 'duration_secs': 0.184387} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.528150] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 856.528421] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-08d8fa3f-a729-4aa3-85c2-19f902f29e0a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.531750] env[62460]: DEBUG oslo_vmware.api [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 856.531750] env[62460]: value = "task-1313680" [ 856.531750] env[62460]: _type = "Task" [ 856.531750] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.535862] env[62460]: DEBUG oslo_vmware.api [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Waiting for the task: (returnval){ [ 856.535862] env[62460]: value = "task-1313681" [ 856.535862] env[62460]: _type = "Task" [ 856.535862] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.542790] env[62460]: DEBUG oslo_vmware.api [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313680, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.548084] env[62460]: DEBUG oslo_vmware.api [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Task: {'id': task-1313681, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.622649] env[62460]: DEBUG nova.scheduler.client.report [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 856.918035] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7449cc4c-6297-42ca-a865-12b374c68f4b tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Lock "a02b4be1-91b0-4254-8d60-654885e24f6b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 74.731s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.928355] env[62460]: DEBUG oslo_vmware.api [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Task: {'id': task-1313677, 'name': PowerOnVM_Task, 'duration_secs': 0.5442} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.928693] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 856.931936] env[62460]: DEBUG nova.compute.manager [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 856.931936] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acf971e3-2bea-42d2-94f2-e374db2da7e6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.953990] env[62460]: DEBUG nova.compute.manager [None req-5770ec75-799a-4db6-97a5-b98712f6f11d tempest-ServersListShow296Test-180931441 tempest-ServersListShow296Test-180931441-project-member] [instance: 120b7508-dc38-4659-92cb-8b3e813dae2a] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 856.985215] env[62460]: DEBUG nova.network.neutron [None req-44d737cd-9ee5-48a0-baac-6819166b7570 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Successfully updated port: d1397bb8-e534-405e-9921-e3e7e6914f81 {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 857.013987] env[62460]: DEBUG oslo_vmware.api [None req-a0a3bf88-42e5-4bc9-9abb-b76cbcd709f3 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313679, 'name': PowerOffVM_Task, 'duration_secs': 0.362127} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.014499] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0a3bf88-42e5-4bc9-9abb-b76cbcd709f3 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 857.014952] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a0a3bf88-42e5-4bc9-9abb-b76cbcd709f3 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 857.015147] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e6368bef-8696-437c-be5f-674b6e60a39e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.047145] env[62460]: DEBUG oslo_vmware.api [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Task: {'id': task-1313681, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.050239] env[62460]: DEBUG oslo_vmware.api [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313680, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.093895} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.050552] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 857.051278] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc5c55ab-451f-45aa-9074-62911b756248 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.073839] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] 896f8ff4-e45f-4403-a727-03ee25e58609/896f8ff4-e45f-4403-a727-03ee25e58609.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 857.074589] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-82145786-1567-40ff-ad78-73bde57b2ca7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.089931] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a0a3bf88-42e5-4bc9-9abb-b76cbcd709f3 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 857.090166] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a0a3bf88-42e5-4bc9-9abb-b76cbcd709f3 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] Deleting contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 857.090404] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-a0a3bf88-42e5-4bc9-9abb-b76cbcd709f3 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Deleting the datastore file [datastore1] 16d40829-ba6e-4193-98b5-fff94c066bcb {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 857.091127] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d105acc9-ce93-425a-922c-afd93823e629 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.096098] env[62460]: DEBUG oslo_vmware.api [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 857.096098] env[62460]: value = "task-1313683" [ 857.096098] env[62460]: _type = "Task" [ 857.096098] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.097429] env[62460]: DEBUG oslo_vmware.api [None req-a0a3bf88-42e5-4bc9-9abb-b76cbcd709f3 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for the task: (returnval){ [ 857.097429] env[62460]: value = "task-1313684" [ 857.097429] env[62460]: _type = "Task" [ 857.097429] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.107437] env[62460]: DEBUG oslo_vmware.api [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313683, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.110290] env[62460]: DEBUG oslo_vmware.api [None req-a0a3bf88-42e5-4bc9-9abb-b76cbcd709f3 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313684, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.111031] env[62460]: DEBUG nova.network.neutron [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Successfully updated port: dbacff36-64b1-4fc7-b0a0-dc78fcb72ac2 {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 857.128315] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.514s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.128861] env[62460]: DEBUG nova.compute.manager [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 857.131209] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 17.409s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.450701] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.477587] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5770ec75-799a-4db6-97a5-b98712f6f11d tempest-ServersListShow296Test-180931441 tempest-ServersListShow296Test-180931441-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.487525] env[62460]: DEBUG oslo_concurrency.lockutils [None req-44d737cd-9ee5-48a0-baac-6819166b7570 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "refresh_cache-7e6ff902-4a04-43d5-9014-38c4ec88efc4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.487736] env[62460]: DEBUG oslo_concurrency.lockutils [None req-44d737cd-9ee5-48a0-baac-6819166b7570 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquired lock "refresh_cache-7e6ff902-4a04-43d5-9014-38c4ec88efc4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.487924] env[62460]: DEBUG nova.network.neutron [None req-44d737cd-9ee5-48a0-baac-6819166b7570 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 857.548044] env[62460]: DEBUG oslo_vmware.api [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Task: {'id': task-1313681, 'name': PowerOnVM_Task, 'duration_secs': 0.830477} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.548426] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 857.548704] env[62460]: INFO nova.compute.manager [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] Took 8.71 seconds to spawn the instance on the hypervisor. [ 857.548839] env[62460]: DEBUG nova.compute.manager [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 857.549620] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cccad7e0-254f-4ba6-9c9e-a0301484f4d2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.612124] env[62460]: DEBUG oslo_vmware.api [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313683, 'name': ReconfigVM_Task, 'duration_secs': 0.495542} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.612387] env[62460]: DEBUG oslo_vmware.api [None req-a0a3bf88-42e5-4bc9-9abb-b76cbcd709f3 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313684, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.345308} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.613401] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Reconfigured VM instance instance-0000004a to attach disk [datastore1] 896f8ff4-e45f-4403-a727-03ee25e58609/896f8ff4-e45f-4403-a727-03ee25e58609.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 857.614303] env[62460]: DEBUG oslo_concurrency.lockutils [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Acquiring lock "refresh_cache-f1921907-f69c-49a3-896c-a4b12b526ffa" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.614443] env[62460]: DEBUG oslo_concurrency.lockutils [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Acquired lock "refresh_cache-f1921907-f69c-49a3-896c-a4b12b526ffa" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.614590] env[62460]: DEBUG nova.network.neutron [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 857.615652] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-a0a3bf88-42e5-4bc9-9abb-b76cbcd709f3 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 857.615889] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a0a3bf88-42e5-4bc9-9abb-b76cbcd709f3 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] Deleted contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 857.616176] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a0a3bf88-42e5-4bc9-9abb-b76cbcd709f3 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 857.616369] env[62460]: INFO nova.compute.manager [None req-a0a3bf88-42e5-4bc9-9abb-b76cbcd709f3 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] Took 1.13 seconds to destroy the instance on the hypervisor. [ 857.616616] env[62460]: DEBUG oslo.service.loopingcall [None req-a0a3bf88-42e5-4bc9-9abb-b76cbcd709f3 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 857.617118] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b80519d8-ed07-40ee-86bb-75c17577f712 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.618858] env[62460]: DEBUG nova.compute.manager [-] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 857.618963] env[62460]: DEBUG nova.network.neutron [-] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 857.626411] env[62460]: DEBUG oslo_vmware.api [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 857.626411] env[62460]: value = "task-1313685" [ 857.626411] env[62460]: _type = "Task" [ 857.626411] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.642782] env[62460]: DEBUG nova.compute.utils [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 857.644228] env[62460]: DEBUG oslo_vmware.api [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313685, 'name': Rename_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.644816] env[62460]: DEBUG nova.compute.manager [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 857.644984] env[62460]: DEBUG nova.network.neutron [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 857.673668] env[62460]: DEBUG nova.compute.manager [req-7523c045-001a-4c14-a1a6-a52ce9bdbb47 req-488be220-6285-4c12-8dba-1f44eaced463 service nova] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Received event network-vif-plugged-d1397bb8-e534-405e-9921-e3e7e6914f81 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 857.673999] env[62460]: DEBUG oslo_concurrency.lockutils [req-7523c045-001a-4c14-a1a6-a52ce9bdbb47 req-488be220-6285-4c12-8dba-1f44eaced463 service nova] Acquiring lock "7e6ff902-4a04-43d5-9014-38c4ec88efc4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.674322] env[62460]: DEBUG oslo_concurrency.lockutils [req-7523c045-001a-4c14-a1a6-a52ce9bdbb47 req-488be220-6285-4c12-8dba-1f44eaced463 service nova] Lock "7e6ff902-4a04-43d5-9014-38c4ec88efc4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.674598] env[62460]: DEBUG oslo_concurrency.lockutils [req-7523c045-001a-4c14-a1a6-a52ce9bdbb47 req-488be220-6285-4c12-8dba-1f44eaced463 service nova] Lock "7e6ff902-4a04-43d5-9014-38c4ec88efc4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.674874] env[62460]: DEBUG nova.compute.manager [req-7523c045-001a-4c14-a1a6-a52ce9bdbb47 req-488be220-6285-4c12-8dba-1f44eaced463 service nova] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] No waiting events found dispatching network-vif-plugged-d1397bb8-e534-405e-9921-e3e7e6914f81 {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 857.675174] env[62460]: WARNING nova.compute.manager [req-7523c045-001a-4c14-a1a6-a52ce9bdbb47 req-488be220-6285-4c12-8dba-1f44eaced463 service nova] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Received unexpected event network-vif-plugged-d1397bb8-e534-405e-9921-e3e7e6914f81 for instance with vm_state active and task_state None. [ 857.689171] env[62460]: DEBUG nova.policy [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dccf00a6951749a791878e716ccf15fd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b7482bca612d4745a2b540c320c07ac7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 857.747683] env[62460]: DEBUG nova.compute.manager [req-f04af106-f30a-47a9-b621-1d4eca5549ee req-3689f700-3e92-48e3-a78f-8af892f0a155 service nova] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Received event network-vif-plugged-dbacff36-64b1-4fc7-b0a0-dc78fcb72ac2 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 857.747894] env[62460]: DEBUG oslo_concurrency.lockutils [req-f04af106-f30a-47a9-b621-1d4eca5549ee req-3689f700-3e92-48e3-a78f-8af892f0a155 service nova] Acquiring lock "f1921907-f69c-49a3-896c-a4b12b526ffa-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.748126] env[62460]: DEBUG oslo_concurrency.lockutils [req-f04af106-f30a-47a9-b621-1d4eca5549ee req-3689f700-3e92-48e3-a78f-8af892f0a155 service nova] Lock "f1921907-f69c-49a3-896c-a4b12b526ffa-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.748324] env[62460]: DEBUG oslo_concurrency.lockutils [req-f04af106-f30a-47a9-b621-1d4eca5549ee req-3689f700-3e92-48e3-a78f-8af892f0a155 service nova] Lock "f1921907-f69c-49a3-896c-a4b12b526ffa-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.748515] env[62460]: DEBUG nova.compute.manager [req-f04af106-f30a-47a9-b621-1d4eca5549ee req-3689f700-3e92-48e3-a78f-8af892f0a155 service nova] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] No waiting events found dispatching network-vif-plugged-dbacff36-64b1-4fc7-b0a0-dc78fcb72ac2 {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 857.748691] env[62460]: WARNING nova.compute.manager [req-f04af106-f30a-47a9-b621-1d4eca5549ee req-3689f700-3e92-48e3-a78f-8af892f0a155 service nova] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Received unexpected event network-vif-plugged-dbacff36-64b1-4fc7-b0a0-dc78fcb72ac2 for instance with vm_state building and task_state spawning. [ 857.975546] env[62460]: DEBUG nova.compute.manager [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 857.976181] env[62460]: DEBUG nova.virt.hardware [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 857.976441] env[62460]: DEBUG nova.virt.hardware [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 857.976608] env[62460]: DEBUG nova.virt.hardware [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 857.976796] env[62460]: DEBUG nova.virt.hardware [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 857.976984] env[62460]: DEBUG nova.virt.hardware [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 857.977121] env[62460]: DEBUG nova.virt.hardware [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 857.977329] env[62460]: DEBUG nova.virt.hardware [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 857.977496] env[62460]: DEBUG nova.virt.hardware [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 857.977670] env[62460]: DEBUG nova.virt.hardware [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 857.977840] env[62460]: DEBUG nova.virt.hardware [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 857.978058] env[62460]: DEBUG nova.virt.hardware [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 857.978972] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbe1e2c2-d700-4639-b4d8-39e29f79d26d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.987433] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc39a342-db56-41aa-90d0-fec56eeb784c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.029314] env[62460]: WARNING nova.network.neutron [None req-44d737cd-9ee5-48a0-baac-6819166b7570 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] d3136e32-ad55-4b73-835c-8fa5f0480767 already exists in list: networks containing: ['d3136e32-ad55-4b73-835c-8fa5f0480767']. ignoring it [ 858.029529] env[62460]: WARNING nova.network.neutron [None req-44d737cd-9ee5-48a0-baac-6819166b7570 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] d3136e32-ad55-4b73-835c-8fa5f0480767 already exists in list: networks containing: ['d3136e32-ad55-4b73-835c-8fa5f0480767']. ignoring it [ 858.053647] env[62460]: DEBUG nova.network.neutron [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] Successfully created port: 948f172e-f4a4-484c-b845-ccb61e37c1b2 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 858.067092] env[62460]: INFO nova.compute.manager [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] Took 27.03 seconds to build instance. [ 858.136874] env[62460]: DEBUG oslo_vmware.api [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313685, 'name': Rename_Task, 'duration_secs': 0.238516} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.137591] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 858.137591] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-54cacecf-5fd0-4e33-9e2c-c8080378c9d8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.144811] env[62460]: DEBUG oslo_vmware.api [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 858.144811] env[62460]: value = "task-1313686" [ 858.144811] env[62460]: _type = "Task" [ 858.144811] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.145461] env[62460]: DEBUG nova.compute.manager [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 858.170166] env[62460]: DEBUG oslo_vmware.api [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313686, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.171103] env[62460]: DEBUG nova.network.neutron [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 858.182975] env[62460]: WARNING nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 1f318a64-2c38-470b-8fae-4ba4543a5681 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 858.183189] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 7e6ff902-4a04-43d5-9014-38c4ec88efc4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 858.183345] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 72e91f8e-0619-464c-b9bc-d6a14be42cb8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 858.183500] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 3ada3516-3147-4566-a46a-1cb29cf880d0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 858.183678] env[62460]: WARNING nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance fde12685-d3b9-46a0-8931-25b904d4f21e is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 858.183901] env[62460]: WARNING nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance b76f37a0-91d0-4a01-9d95-9c6586081175 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 858.184137] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 5b6e8205-003e-49c0-a73d-be2e032a8272 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 858.184356] env[62460]: WARNING nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 858.184509] env[62460]: WARNING nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 0269dc64-d2b1-43c5-bdf7-11d97e534819 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 858.184610] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 9d1a5830-f3c0-4d18-9338-16f7b6962c6a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 858.184742] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance d7d180f7-e1a8-46c8-ba8e-ca50dac474cc actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 858.184901] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 16d40829-ba6e-4193-98b5-fff94c066bcb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 858.185103] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance a02b4be1-91b0-4254-8d60-654885e24f6b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 858.185246] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 49477d35-92ea-4f9d-8333-fc23144b7dfc actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 858.185408] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 896f8ff4-e45f-4403-a727-03ee25e58609 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 858.185576] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance f1921907-f69c-49a3-896c-a4b12b526ffa actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 858.185742] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 0f5457a0-c918-4407-9ba9-bfeeb949c6a7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 858.294887] env[62460]: DEBUG nova.compute.manager [None req-7ec877fc-e0a0-40d7-aa37-b25eed5a9144 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 858.295810] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbfe29a2-807e-4d4c-aeb5-a59bc4da6223 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.376513] env[62460]: DEBUG nova.network.neutron [-] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 858.398491] env[62460]: DEBUG nova.network.neutron [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Updating instance_info_cache with network_info: [{"id": "dbacff36-64b1-4fc7-b0a0-dc78fcb72ac2", "address": "fa:16:3e:5e:f0:fa", "network": {"id": "828c25a7-673d-4b6c-ad50-7317affa9ded", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-80559771-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a10038b038254e5d9543018e39a56992", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e39ca24f-7890-4cdf-8dab-ecab218bb063", "external-id": "nsx-vlan-transportzone-890", "segmentation_id": 890, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdbacff36-64", "ovs_interfaceid": "dbacff36-64b1-4fc7-b0a0-dc78fcb72ac2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 858.518784] env[62460]: DEBUG nova.network.neutron [None req-44d737cd-9ee5-48a0-baac-6819166b7570 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Updating instance_info_cache with network_info: [{"id": "ba38cbb0-62cf-47cb-a238-87ebc03de250", "address": "fa:16:3e:b4:93:18", "network": {"id": "d3136e32-ad55-4b73-835c-8fa5f0480767", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1511414968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53fa06008e9f43488362895e7a143700", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "054fcd1e-638e-425a-a1de-78cb188ae026", "external-id": "nsx-vlan-transportzone-658", "segmentation_id": 658, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba38cbb0-62", "ovs_interfaceid": "ba38cbb0-62cf-47cb-a238-87ebc03de250", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "bf0f715f-7988-45f9-81b6-c92688ff60c7", "address": "fa:16:3e:84:3c:55", "network": {"id": "d3136e32-ad55-4b73-835c-8fa5f0480767", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1511414968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53fa06008e9f43488362895e7a143700", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "054fcd1e-638e-425a-a1de-78cb188ae026", "external-id": "nsx-vlan-transportzone-658", "segmentation_id": 658, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf0f715f-79", "ovs_interfaceid": "bf0f715f-7988-45f9-81b6-c92688ff60c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "d1397bb8-e534-405e-9921-e3e7e6914f81", "address": "fa:16:3e:2e:b4:bd", "network": {"id": "d3136e32-ad55-4b73-835c-8fa5f0480767", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1511414968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53fa06008e9f43488362895e7a143700", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "054fcd1e-638e-425a-a1de-78cb188ae026", "external-id": "nsx-vlan-transportzone-658", "segmentation_id": 658, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd1397bb8-e5", "ovs_interfaceid": "d1397bb8-e534-405e-9921-e3e7e6914f81", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 858.569672] env[62460]: DEBUG oslo_concurrency.lockutils [None req-144a0602-4e3a-4e05-b47a-43bc5c1a4488 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Lock "49477d35-92ea-4f9d-8333-fc23144b7dfc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 61.413s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.655018] env[62460]: DEBUG oslo_vmware.api [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313686, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.689471] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 3e10af70-db52-4d4a-bb92-821a05dcbab6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 858.807822] env[62460]: INFO nova.compute.manager [None req-7ec877fc-e0a0-40d7-aa37-b25eed5a9144 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] instance snapshotting [ 858.811357] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95f6a2be-1b17-441e-8eae-d8ce0d92f466 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.843056] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0ff4c3e-c758-4edf-98e4-42dc27d2cf46 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.878276] env[62460]: INFO nova.compute.manager [-] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] Took 1.26 seconds to deallocate network for instance. [ 858.901682] env[62460]: DEBUG oslo_concurrency.lockutils [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Releasing lock "refresh_cache-f1921907-f69c-49a3-896c-a4b12b526ffa" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.902064] env[62460]: DEBUG nova.compute.manager [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Instance network_info: |[{"id": "dbacff36-64b1-4fc7-b0a0-dc78fcb72ac2", "address": "fa:16:3e:5e:f0:fa", "network": {"id": "828c25a7-673d-4b6c-ad50-7317affa9ded", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-80559771-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a10038b038254e5d9543018e39a56992", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e39ca24f-7890-4cdf-8dab-ecab218bb063", "external-id": "nsx-vlan-transportzone-890", "segmentation_id": 890, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdbacff36-64", "ovs_interfaceid": "dbacff36-64b1-4fc7-b0a0-dc78fcb72ac2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 858.902799] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5e:f0:fa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e39ca24f-7890-4cdf-8dab-ecab218bb063', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dbacff36-64b1-4fc7-b0a0-dc78fcb72ac2', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 858.910585] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Creating folder: Project (a10038b038254e5d9543018e39a56992). Parent ref: group-v281134. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 858.911801] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6217bee1-517c-48ff-992d-aa84e46edb83 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.924968] env[62460]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 858.925192] env[62460]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=62460) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 858.925875] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Folder already exists: Project (a10038b038254e5d9543018e39a56992). Parent ref: group-v281134. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 858.926122] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Creating folder: Instances. Parent ref: group-v281158. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 858.926385] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-aa10cdd0-9b89-4655-9c73-5b4ca2cebf4d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.937966] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Created folder: Instances in parent group-v281158. [ 858.938268] env[62460]: DEBUG oslo.service.loopingcall [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 858.938512] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 858.939116] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d6abad2e-0ece-4d4f-a7f7-93b19c5e2ab4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.961938] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 858.961938] env[62460]: value = "task-1313689" [ 858.961938] env[62460]: _type = "Task" [ 858.961938] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.970296] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313689, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.021804] env[62460]: DEBUG oslo_concurrency.lockutils [None req-44d737cd-9ee5-48a0-baac-6819166b7570 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Releasing lock "refresh_cache-7e6ff902-4a04-43d5-9014-38c4ec88efc4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.022684] env[62460]: DEBUG oslo_concurrency.lockutils [None req-44d737cd-9ee5-48a0-baac-6819166b7570 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "7e6ff902-4a04-43d5-9014-38c4ec88efc4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.022904] env[62460]: DEBUG oslo_concurrency.lockutils [None req-44d737cd-9ee5-48a0-baac-6819166b7570 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquired lock "7e6ff902-4a04-43d5-9014-38c4ec88efc4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.024063] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c61cc17a-2803-4d5b-ba03-6cc5d4eb8aab {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.044191] env[62460]: DEBUG nova.virt.hardware [None req-44d737cd-9ee5-48a0-baac-6819166b7570 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 859.044528] env[62460]: DEBUG nova.virt.hardware [None req-44d737cd-9ee5-48a0-baac-6819166b7570 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 859.044741] env[62460]: DEBUG nova.virt.hardware [None req-44d737cd-9ee5-48a0-baac-6819166b7570 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 859.045054] env[62460]: DEBUG nova.virt.hardware [None req-44d737cd-9ee5-48a0-baac-6819166b7570 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 859.045192] env[62460]: DEBUG nova.virt.hardware [None req-44d737cd-9ee5-48a0-baac-6819166b7570 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 859.045431] env[62460]: DEBUG nova.virt.hardware [None req-44d737cd-9ee5-48a0-baac-6819166b7570 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 859.045665] env[62460]: DEBUG nova.virt.hardware [None req-44d737cd-9ee5-48a0-baac-6819166b7570 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 859.045838] env[62460]: DEBUG nova.virt.hardware [None req-44d737cd-9ee5-48a0-baac-6819166b7570 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 859.046053] env[62460]: DEBUG nova.virt.hardware [None req-44d737cd-9ee5-48a0-baac-6819166b7570 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 859.046381] env[62460]: DEBUG nova.virt.hardware [None req-44d737cd-9ee5-48a0-baac-6819166b7570 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 859.046461] env[62460]: DEBUG nova.virt.hardware [None req-44d737cd-9ee5-48a0-baac-6819166b7570 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 859.054086] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-44d737cd-9ee5-48a0-baac-6819166b7570 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Reconfiguring VM to attach interface {{(pid=62460) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 859.054231] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-61d05114-2e80-4f95-96e6-3ec88a3497fc {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.072685] env[62460]: DEBUG oslo_vmware.api [None req-44d737cd-9ee5-48a0-baac-6819166b7570 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Waiting for the task: (returnval){ [ 859.072685] env[62460]: value = "task-1313690" [ 859.072685] env[62460]: _type = "Task" [ 859.072685] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.081561] env[62460]: DEBUG oslo_vmware.api [None req-44d737cd-9ee5-48a0-baac-6819166b7570 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313690, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.156031] env[62460]: DEBUG oslo_vmware.api [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313686, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.162626] env[62460]: DEBUG nova.compute.manager [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 859.192979] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance fd03509c-5471-48bb-8150-d29531dca848 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 859.199352] env[62460]: DEBUG nova.virt.hardware [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 859.199681] env[62460]: DEBUG nova.virt.hardware [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 859.199812] env[62460]: DEBUG nova.virt.hardware [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 859.200014] env[62460]: DEBUG nova.virt.hardware [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 859.200193] env[62460]: DEBUG nova.virt.hardware [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 859.200428] env[62460]: DEBUG nova.virt.hardware [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 859.200717] env[62460]: DEBUG nova.virt.hardware [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 859.200903] env[62460]: DEBUG nova.virt.hardware [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 859.201517] env[62460]: DEBUG nova.virt.hardware [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 859.201788] env[62460]: DEBUG nova.virt.hardware [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 859.202127] env[62460]: DEBUG nova.virt.hardware [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 859.203404] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1662c31-637b-44d5-9585-380e357d3fbd {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.215832] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc3fd501-cc92-4c18-8629-121fd5e7a7f6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.359971] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7ec877fc-e0a0-40d7-aa37-b25eed5a9144 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Creating Snapshot of the VM instance {{(pid=62460) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 859.360884] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-96a4ef79-c446-4dfe-8513-a803ed1fc512 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.369631] env[62460]: DEBUG oslo_vmware.api [None req-7ec877fc-e0a0-40d7-aa37-b25eed5a9144 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Waiting for the task: (returnval){ [ 859.369631] env[62460]: value = "task-1313691" [ 859.369631] env[62460]: _type = "Task" [ 859.369631] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.381266] env[62460]: DEBUG oslo_vmware.api [None req-7ec877fc-e0a0-40d7-aa37-b25eed5a9144 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Task: {'id': task-1313691, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.389566] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a0a3bf88-42e5-4bc9-9abb-b76cbcd709f3 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.478070] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313689, 'name': CreateVM_Task, 'duration_secs': 0.36292} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.478070] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 859.478070] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'mount_device': '/dev/sda', 'device_type': None, 'attachment_id': '857b0225-8e87-4ba9-b6c1-d9ab66b4754b', 'guest_format': None, 'disk_bus': None, 'boot_index': 0, 'delete_on_termination': True, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281168', 'volume_id': '2cf093c7-9278-4f2d-a385-8775acc7dae8', 'name': 'volume-2cf093c7-9278-4f2d-a385-8775acc7dae8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f1921907-f69c-49a3-896c-a4b12b526ffa', 'attached_at': '', 'detached_at': '', 'volume_id': '2cf093c7-9278-4f2d-a385-8775acc7dae8', 'serial': '2cf093c7-9278-4f2d-a385-8775acc7dae8'}, 'volume_type': None}], 'swap': None} {{(pid=62460) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 859.478070] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Root volume attach. Driver type: vmdk {{(pid=62460) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 859.479973] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5961ea5c-9f45-46da-ba82-8b3705a060e6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.493302] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2720686-038a-40ad-92b9-8aedb793cea8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.501599] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09b6cebf-192c-4400-a4df-12d33bee07e8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.509822] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-2d2a634b-01cc-4d84-85af-903b8856151f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.517499] env[62460]: DEBUG oslo_vmware.api [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Waiting for the task: (returnval){ [ 859.517499] env[62460]: value = "task-1313692" [ 859.517499] env[62460]: _type = "Task" [ 859.517499] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.525490] env[62460]: DEBUG oslo_vmware.api [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Task: {'id': task-1313692, 'name': RelocateVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.583291] env[62460]: DEBUG oslo_vmware.api [None req-44d737cd-9ee5-48a0-baac-6819166b7570 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313690, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.656635] env[62460]: DEBUG oslo_vmware.api [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313686, 'name': PowerOnVM_Task, 'duration_secs': 1.238232} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.657060] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 859.657347] env[62460]: INFO nova.compute.manager [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Took 8.42 seconds to spawn the instance on the hypervisor. [ 859.657566] env[62460]: DEBUG nova.compute.manager [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 859.658544] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25e4d226-eabb-4888-ac99-7f52a38d500f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.708646] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a8b2d4df-1735-4f45-bc97-65d17e36a285 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Acquiring lock "9d1a5830-f3c0-4d18-9338-16f7b6962c6a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.710019] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a8b2d4df-1735-4f45-bc97-65d17e36a285 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Lock "9d1a5830-f3c0-4d18-9338-16f7b6962c6a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.710019] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a8b2d4df-1735-4f45-bc97-65d17e36a285 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Acquiring lock "9d1a5830-f3c0-4d18-9338-16f7b6962c6a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.710019] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a8b2d4df-1735-4f45-bc97-65d17e36a285 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Lock "9d1a5830-f3c0-4d18-9338-16f7b6962c6a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.710019] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a8b2d4df-1735-4f45-bc97-65d17e36a285 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Lock "9d1a5830-f3c0-4d18-9338-16f7b6962c6a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.712617] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 120b7508-dc38-4659-92cb-8b3e813dae2a has been deleted (perhaps locally). Deleting allocations that remained for this instance against this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1750}} [ 859.717579] env[62460]: INFO nova.compute.manager [None req-a8b2d4df-1735-4f45-bc97-65d17e36a285 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Terminating instance [ 859.719998] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a8b2d4df-1735-4f45-bc97-65d17e36a285 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Acquiring lock "refresh_cache-9d1a5830-f3c0-4d18-9338-16f7b6962c6a" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.720308] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a8b2d4df-1735-4f45-bc97-65d17e36a285 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Acquired lock "refresh_cache-9d1a5830-f3c0-4d18-9338-16f7b6962c6a" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.720378] env[62460]: DEBUG nova.network.neutron [None req-a8b2d4df-1735-4f45-bc97-65d17e36a285 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 859.733788] env[62460]: DEBUG nova.scheduler.client.report [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Cannot delete allocation for 120b7508-dc38-4659-92cb-8b3e813dae2a consumer in placement as consumer does not exist {{(pid=62460) delete_allocation_for_instance /opt/stack/nova/nova/scheduler/client/report.py:2199}} [ 859.734056] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Total usable vcpus: 48, total allocated vcpus: 12 {{(pid=62460) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 859.734227] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2816MB phys_disk=200GB used_disk=11GB total_vcpus=48 used_vcpus=12 pci_stats=[] {{(pid=62460) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 859.884761] env[62460]: DEBUG oslo_vmware.api [None req-7ec877fc-e0a0-40d7-aa37-b25eed5a9144 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Task: {'id': task-1313691, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.901976] env[62460]: DEBUG nova.network.neutron [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] Successfully updated port: 948f172e-f4a4-484c-b845-ccb61e37c1b2 {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 860.014617] env[62460]: DEBUG nova.compute.manager [req-8486fc4d-0c1c-48ed-b6f7-f3202c1499ab req-ddc4dd62-e42c-483f-9929-677bc9b17764 service nova] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Received event network-changed-d1397bb8-e534-405e-9921-e3e7e6914f81 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 860.014788] env[62460]: DEBUG nova.compute.manager [req-8486fc4d-0c1c-48ed-b6f7-f3202c1499ab req-ddc4dd62-e42c-483f-9929-677bc9b17764 service nova] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Refreshing instance network info cache due to event network-changed-d1397bb8-e534-405e-9921-e3e7e6914f81. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 860.015582] env[62460]: DEBUG oslo_concurrency.lockutils [req-8486fc4d-0c1c-48ed-b6f7-f3202c1499ab req-ddc4dd62-e42c-483f-9929-677bc9b17764 service nova] Acquiring lock "refresh_cache-7e6ff902-4a04-43d5-9014-38c4ec88efc4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.015582] env[62460]: DEBUG oslo_concurrency.lockutils [req-8486fc4d-0c1c-48ed-b6f7-f3202c1499ab req-ddc4dd62-e42c-483f-9929-677bc9b17764 service nova] Acquired lock "refresh_cache-7e6ff902-4a04-43d5-9014-38c4ec88efc4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.015582] env[62460]: DEBUG nova.network.neutron [req-8486fc4d-0c1c-48ed-b6f7-f3202c1499ab req-ddc4dd62-e42c-483f-9929-677bc9b17764 service nova] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Refreshing network info cache for port d1397bb8-e534-405e-9921-e3e7e6914f81 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 860.032191] env[62460]: DEBUG oslo_vmware.api [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Task: {'id': task-1313692, 'name': RelocateVM_Task} progress is 20%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.048097] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1797a7e7-d305-4274-8a5e-afbe37fedb40 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.056059] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79cdf63d-cf39-4967-9dfe-a061db6e9596 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.092353] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9893d1ec-3b4a-4d08-b963-b9d7b26b6a75 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.096312] env[62460]: DEBUG nova.compute.manager [req-90285a8b-66da-44cc-b28e-f22f45bf6300 req-a272e1ad-b526-4bbe-afeb-5b28cce4e707 service nova] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Received event network-changed-dbacff36-64b1-4fc7-b0a0-dc78fcb72ac2 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 860.096581] env[62460]: DEBUG nova.compute.manager [req-90285a8b-66da-44cc-b28e-f22f45bf6300 req-a272e1ad-b526-4bbe-afeb-5b28cce4e707 service nova] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Refreshing instance network info cache due to event network-changed-dbacff36-64b1-4fc7-b0a0-dc78fcb72ac2. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 860.096826] env[62460]: DEBUG oslo_concurrency.lockutils [req-90285a8b-66da-44cc-b28e-f22f45bf6300 req-a272e1ad-b526-4bbe-afeb-5b28cce4e707 service nova] Acquiring lock "refresh_cache-f1921907-f69c-49a3-896c-a4b12b526ffa" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.096979] env[62460]: DEBUG oslo_concurrency.lockutils [req-90285a8b-66da-44cc-b28e-f22f45bf6300 req-a272e1ad-b526-4bbe-afeb-5b28cce4e707 service nova] Acquired lock "refresh_cache-f1921907-f69c-49a3-896c-a4b12b526ffa" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.097208] env[62460]: DEBUG nova.network.neutron [req-90285a8b-66da-44cc-b28e-f22f45bf6300 req-a272e1ad-b526-4bbe-afeb-5b28cce4e707 service nova] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Refreshing network info cache for port dbacff36-64b1-4fc7-b0a0-dc78fcb72ac2 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 860.105189] env[62460]: DEBUG oslo_vmware.api [None req-44d737cd-9ee5-48a0-baac-6819166b7570 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313690, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.107969] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d417ed85-fb91-4e26-8934-ecccd2e6a497 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.123454] env[62460]: DEBUG nova.compute.provider_tree [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 860.182328] env[62460]: INFO nova.compute.manager [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Took 28.77 seconds to build instance. [ 860.239358] env[62460]: DEBUG nova.network.neutron [None req-a8b2d4df-1735-4f45-bc97-65d17e36a285 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 860.331089] env[62460]: DEBUG nova.network.neutron [None req-a8b2d4df-1735-4f45-bc97-65d17e36a285 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.381791] env[62460]: DEBUG oslo_vmware.api [None req-7ec877fc-e0a0-40d7-aa37-b25eed5a9144 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Task: {'id': task-1313691, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.406090] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Acquiring lock "refresh_cache-0f5457a0-c918-4407-9ba9-bfeeb949c6a7" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.406277] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Acquired lock "refresh_cache-0f5457a0-c918-4407-9ba9-bfeeb949c6a7" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.406484] env[62460]: DEBUG nova.network.neutron [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 860.530180] env[62460]: DEBUG oslo_vmware.api [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Task: {'id': task-1313692, 'name': RelocateVM_Task, 'duration_secs': 0.553988} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.531012] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Volume attach. Driver type: vmdk {{(pid=62460) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 860.531012] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281168', 'volume_id': '2cf093c7-9278-4f2d-a385-8775acc7dae8', 'name': 'volume-2cf093c7-9278-4f2d-a385-8775acc7dae8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f1921907-f69c-49a3-896c-a4b12b526ffa', 'attached_at': '', 'detached_at': '', 'volume_id': '2cf093c7-9278-4f2d-a385-8775acc7dae8', 'serial': '2cf093c7-9278-4f2d-a385-8775acc7dae8'} {{(pid=62460) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 860.531404] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dafe95c-749b-47d6-8c1a-a9c8d43ef586 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.547429] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c76c3d61-d8f6-4fc8-bf99-2feebd1a1e9b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.571322] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Reconfiguring VM instance instance-0000004b to attach disk [datastore2] volume-2cf093c7-9278-4f2d-a385-8775acc7dae8/volume-2cf093c7-9278-4f2d-a385-8775acc7dae8.vmdk or device None with type thin {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 860.573812] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-39a1fa47-bf83-40f7-87d1-a30c1aa25469 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.592522] env[62460]: DEBUG oslo_vmware.api [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Waiting for the task: (returnval){ [ 860.592522] env[62460]: value = "task-1313693" [ 860.592522] env[62460]: _type = "Task" [ 860.592522] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.595610] env[62460]: DEBUG oslo_vmware.api [None req-44d737cd-9ee5-48a0-baac-6819166b7570 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313690, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.604596] env[62460]: DEBUG oslo_vmware.api [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Task: {'id': task-1313693, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.627105] env[62460]: DEBUG nova.scheduler.client.report [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 860.684490] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ce2fa57-f132-412d-8514-f2ac1313c3e8 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lock "896f8ff4-e45f-4403-a727-03ee25e58609" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 52.656s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.823387] env[62460]: DEBUG nova.network.neutron [req-8486fc4d-0c1c-48ed-b6f7-f3202c1499ab req-ddc4dd62-e42c-483f-9929-677bc9b17764 service nova] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Updated VIF entry in instance network info cache for port d1397bb8-e534-405e-9921-e3e7e6914f81. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 860.824237] env[62460]: DEBUG nova.network.neutron [req-8486fc4d-0c1c-48ed-b6f7-f3202c1499ab req-ddc4dd62-e42c-483f-9929-677bc9b17764 service nova] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Updating instance_info_cache with network_info: [{"id": "ba38cbb0-62cf-47cb-a238-87ebc03de250", "address": "fa:16:3e:b4:93:18", "network": {"id": "d3136e32-ad55-4b73-835c-8fa5f0480767", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1511414968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53fa06008e9f43488362895e7a143700", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "054fcd1e-638e-425a-a1de-78cb188ae026", "external-id": "nsx-vlan-transportzone-658", "segmentation_id": 658, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba38cbb0-62", "ovs_interfaceid": "ba38cbb0-62cf-47cb-a238-87ebc03de250", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "bf0f715f-7988-45f9-81b6-c92688ff60c7", "address": "fa:16:3e:84:3c:55", "network": {"id": "d3136e32-ad55-4b73-835c-8fa5f0480767", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1511414968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53fa06008e9f43488362895e7a143700", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "054fcd1e-638e-425a-a1de-78cb188ae026", "external-id": "nsx-vlan-transportzone-658", "segmentation_id": 658, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf0f715f-79", "ovs_interfaceid": "bf0f715f-7988-45f9-81b6-c92688ff60c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "d1397bb8-e534-405e-9921-e3e7e6914f81", "address": "fa:16:3e:2e:b4:bd", "network": {"id": "d3136e32-ad55-4b73-835c-8fa5f0480767", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1511414968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53fa06008e9f43488362895e7a143700", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "054fcd1e-638e-425a-a1de-78cb188ae026", "external-id": "nsx-vlan-transportzone-658", "segmentation_id": 658, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd1397bb8-e5", "ovs_interfaceid": "d1397bb8-e534-405e-9921-e3e7e6914f81", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.834449] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a8b2d4df-1735-4f45-bc97-65d17e36a285 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Releasing lock "refresh_cache-9d1a5830-f3c0-4d18-9338-16f7b6962c6a" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.834949] env[62460]: DEBUG nova.compute.manager [None req-a8b2d4df-1735-4f45-bc97-65d17e36a285 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 860.834949] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a8b2d4df-1735-4f45-bc97-65d17e36a285 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 860.837917] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63577e30-2631-4382-92bf-70a3ab7f7716 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.842046] env[62460]: DEBUG nova.network.neutron [req-90285a8b-66da-44cc-b28e-f22f45bf6300 req-a272e1ad-b526-4bbe-afeb-5b28cce4e707 service nova] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Updated VIF entry in instance network info cache for port dbacff36-64b1-4fc7-b0a0-dc78fcb72ac2. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 860.842589] env[62460]: DEBUG nova.network.neutron [req-90285a8b-66da-44cc-b28e-f22f45bf6300 req-a272e1ad-b526-4bbe-afeb-5b28cce4e707 service nova] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Updating instance_info_cache with network_info: [{"id": "dbacff36-64b1-4fc7-b0a0-dc78fcb72ac2", "address": "fa:16:3e:5e:f0:fa", "network": {"id": "828c25a7-673d-4b6c-ad50-7317affa9ded", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-80559771-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a10038b038254e5d9543018e39a56992", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e39ca24f-7890-4cdf-8dab-ecab218bb063", "external-id": "nsx-vlan-transportzone-890", "segmentation_id": 890, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdbacff36-64", "ovs_interfaceid": "dbacff36-64b1-4fc7-b0a0-dc78fcb72ac2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.849760] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8b2d4df-1735-4f45-bc97-65d17e36a285 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 860.850088] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f1f64903-9cb4-45f6-acff-b9b13a420f3b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.857407] env[62460]: DEBUG oslo_vmware.api [None req-a8b2d4df-1735-4f45-bc97-65d17e36a285 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Waiting for the task: (returnval){ [ 860.857407] env[62460]: value = "task-1313694" [ 860.857407] env[62460]: _type = "Task" [ 860.857407] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.872183] env[62460]: DEBUG oslo_vmware.api [None req-a8b2d4df-1735-4f45-bc97-65d17e36a285 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Task: {'id': task-1313694, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.885342] env[62460]: DEBUG oslo_vmware.api [None req-7ec877fc-e0a0-40d7-aa37-b25eed5a9144 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Task: {'id': task-1313691, 'name': CreateSnapshot_Task, 'duration_secs': 1.084908} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.885747] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7ec877fc-e0a0-40d7-aa37-b25eed5a9144 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Created Snapshot of the VM instance {{(pid=62460) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 860.886707] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca61154c-6547-4384-9ce3-426baa6663b8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.945607] env[62460]: DEBUG nova.network.neutron [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 861.095050] env[62460]: DEBUG nova.network.neutron [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] Updating instance_info_cache with network_info: [{"id": "948f172e-f4a4-484c-b845-ccb61e37c1b2", "address": "fa:16:3e:d1:66:a7", "network": {"id": "c2c0d035-a856-4110-b92d-8a630ba71d3c", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1252073445-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b7482bca612d4745a2b540c320c07ac7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4b43a78-f49b-4132-ab2e-6e28769a9498", "external-id": "nsx-vlan-transportzone-737", "segmentation_id": 737, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap948f172e-f4", "ovs_interfaceid": "948f172e-f4a4-484c-b845-ccb61e37c1b2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.102361] env[62460]: DEBUG oslo_vmware.api [None req-44d737cd-9ee5-48a0-baac-6819166b7570 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313690, 'name': ReconfigVM_Task, 'duration_secs': 1.812312} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.103146] env[62460]: DEBUG oslo_concurrency.lockutils [None req-44d737cd-9ee5-48a0-baac-6819166b7570 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Releasing lock "7e6ff902-4a04-43d5-9014-38c4ec88efc4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.103381] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-44d737cd-9ee5-48a0-baac-6819166b7570 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Reconfigured VM to attach interface {{(pid=62460) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 861.110794] env[62460]: DEBUG oslo_vmware.api [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Task: {'id': task-1313693, 'name': ReconfigVM_Task, 'duration_secs': 0.316273} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.111762] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Reconfigured VM instance instance-0000004b to attach disk [datastore2] volume-2cf093c7-9278-4f2d-a385-8775acc7dae8/volume-2cf093c7-9278-4f2d-a385-8775acc7dae8.vmdk or device None with type thin {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 861.119297] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0918a188-609e-4216-8825-dad47b8dd074 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.131900] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62460) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 861.132217] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 4.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.133860] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.268s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.135648] env[62460]: INFO nova.compute.claims [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 861.139190] env[62460]: DEBUG oslo_vmware.api [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Waiting for the task: (returnval){ [ 861.139190] env[62460]: value = "task-1313695" [ 861.139190] env[62460]: _type = "Task" [ 861.139190] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.150899] env[62460]: DEBUG oslo_vmware.api [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Task: {'id': task-1313695, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.328463] env[62460]: DEBUG oslo_concurrency.lockutils [req-8486fc4d-0c1c-48ed-b6f7-f3202c1499ab req-ddc4dd62-e42c-483f-9929-677bc9b17764 service nova] Releasing lock "refresh_cache-7e6ff902-4a04-43d5-9014-38c4ec88efc4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.345571] env[62460]: DEBUG oslo_concurrency.lockutils [req-90285a8b-66da-44cc-b28e-f22f45bf6300 req-a272e1ad-b526-4bbe-afeb-5b28cce4e707 service nova] Releasing lock "refresh_cache-f1921907-f69c-49a3-896c-a4b12b526ffa" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.346994] env[62460]: DEBUG nova.compute.manager [req-90285a8b-66da-44cc-b28e-f22f45bf6300 req-a272e1ad-b526-4bbe-afeb-5b28cce4e707 service nova] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] Received event network-vif-deleted-b1e3ec03-310d-4b60-8ead-a10ec31a5584 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 861.370517] env[62460]: DEBUG oslo_vmware.api [None req-a8b2d4df-1735-4f45-bc97-65d17e36a285 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Task: {'id': task-1313694, 'name': PowerOffVM_Task, 'duration_secs': 0.308813} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.370781] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8b2d4df-1735-4f45-bc97-65d17e36a285 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 861.370965] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a8b2d4df-1735-4f45-bc97-65d17e36a285 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 861.371256] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6f6b29e9-5524-4a17-ab66-1222bc07b0ec {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.409023] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7ec877fc-e0a0-40d7-aa37-b25eed5a9144 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Creating linked-clone VM from snapshot {{(pid=62460) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 861.409023] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a8b2d4df-1735-4f45-bc97-65d17e36a285 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 861.409023] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a8b2d4df-1735-4f45-bc97-65d17e36a285 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Deleting contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 861.409023] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8b2d4df-1735-4f45-bc97-65d17e36a285 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Deleting the datastore file [datastore1] 9d1a5830-f3c0-4d18-9338-16f7b6962c6a {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 861.409023] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-fb549bba-b241-4bd5-ae99-0752bfe207a6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.410150] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a1bda567-294a-4ce1-a667-e76375bc8df4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.416654] env[62460]: DEBUG oslo_vmware.api [None req-a8b2d4df-1735-4f45-bc97-65d17e36a285 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Waiting for the task: (returnval){ [ 861.416654] env[62460]: value = "task-1313698" [ 861.416654] env[62460]: _type = "Task" [ 861.416654] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.418322] env[62460]: DEBUG oslo_vmware.api [None req-7ec877fc-e0a0-40d7-aa37-b25eed5a9144 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Waiting for the task: (returnval){ [ 861.418322] env[62460]: value = "task-1313697" [ 861.418322] env[62460]: _type = "Task" [ 861.418322] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.431765] env[62460]: DEBUG oslo_vmware.api [None req-a8b2d4df-1735-4f45-bc97-65d17e36a285 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Task: {'id': task-1313698, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.436233] env[62460]: DEBUG oslo_vmware.api [None req-7ec877fc-e0a0-40d7-aa37-b25eed5a9144 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Task: {'id': task-1313697, 'name': CloneVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.603944] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Releasing lock "refresh_cache-0f5457a0-c918-4407-9ba9-bfeeb949c6a7" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.604378] env[62460]: DEBUG nova.compute.manager [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] Instance network_info: |[{"id": "948f172e-f4a4-484c-b845-ccb61e37c1b2", "address": "fa:16:3e:d1:66:a7", "network": {"id": "c2c0d035-a856-4110-b92d-8a630ba71d3c", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1252073445-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b7482bca612d4745a2b540c320c07ac7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4b43a78-f49b-4132-ab2e-6e28769a9498", "external-id": "nsx-vlan-transportzone-737", "segmentation_id": 737, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap948f172e-f4", "ovs_interfaceid": "948f172e-f4a4-484c-b845-ccb61e37c1b2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 861.604904] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d1:66:a7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd4b43a78-f49b-4132-ab2e-6e28769a9498', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '948f172e-f4a4-484c-b845-ccb61e37c1b2', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 861.612467] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Creating folder: Project (b7482bca612d4745a2b540c320c07ac7). Parent ref: group-v281134. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 861.613771] env[62460]: DEBUG oslo_concurrency.lockutils [None req-44d737cd-9ee5-48a0-baac-6819166b7570 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "interface-7e6ff902-4a04-43d5-9014-38c4ec88efc4-d1397bb8-e534-405e-9921-e3e7e6914f81" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.551s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.614849] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-51ea1077-0eec-4cfb-9f58-8d05dd13738b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.625836] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Created folder: Project (b7482bca612d4745a2b540c320c07ac7) in parent group-v281134. [ 861.626095] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Creating folder: Instances. Parent ref: group-v281216. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 861.626354] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-64175596-9fde-4bec-a241-e04275c67b46 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.636383] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Created folder: Instances in parent group-v281216. [ 861.636770] env[62460]: DEBUG oslo.service.loopingcall [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 861.636889] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 861.637430] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b44995ef-62b5-4c81-be38-a145aed8f342 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.666686] env[62460]: DEBUG oslo_vmware.api [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Task: {'id': task-1313695, 'name': ReconfigVM_Task, 'duration_secs': 0.13462} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.667977] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281168', 'volume_id': '2cf093c7-9278-4f2d-a385-8775acc7dae8', 'name': 'volume-2cf093c7-9278-4f2d-a385-8775acc7dae8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f1921907-f69c-49a3-896c-a4b12b526ffa', 'attached_at': '', 'detached_at': '', 'volume_id': '2cf093c7-9278-4f2d-a385-8775acc7dae8', 'serial': '2cf093c7-9278-4f2d-a385-8775acc7dae8'} {{(pid=62460) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 861.668556] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 861.668556] env[62460]: value = "task-1313701" [ 861.668556] env[62460]: _type = "Task" [ 861.668556] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.669348] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-da83a359-bc4c-40e0-b396-7b515de17133 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.681234] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313701, 'name': CreateVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.681554] env[62460]: DEBUG oslo_vmware.api [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Waiting for the task: (returnval){ [ 861.681554] env[62460]: value = "task-1313702" [ 861.681554] env[62460]: _type = "Task" [ 861.681554] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.690739] env[62460]: DEBUG oslo_vmware.api [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Task: {'id': task-1313702, 'name': Rename_Task} progress is 10%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.936667] env[62460]: DEBUG oslo_vmware.api [None req-7ec877fc-e0a0-40d7-aa37-b25eed5a9144 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Task: {'id': task-1313697, 'name': CloneVM_Task} progress is 94%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.941097] env[62460]: DEBUG oslo_vmware.api [None req-a8b2d4df-1735-4f45-bc97-65d17e36a285 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Task: {'id': task-1313698, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.20074} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.941395] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8b2d4df-1735-4f45-bc97-65d17e36a285 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 861.941601] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a8b2d4df-1735-4f45-bc97-65d17e36a285 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Deleted contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 861.941853] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a8b2d4df-1735-4f45-bc97-65d17e36a285 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 861.942359] env[62460]: INFO nova.compute.manager [None req-a8b2d4df-1735-4f45-bc97-65d17e36a285 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Took 1.11 seconds to destroy the instance on the hypervisor. [ 861.942359] env[62460]: DEBUG oslo.service.loopingcall [None req-a8b2d4df-1735-4f45-bc97-65d17e36a285 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 861.942510] env[62460]: DEBUG nova.compute.manager [-] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 861.942613] env[62460]: DEBUG nova.network.neutron [-] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 861.962463] env[62460]: DEBUG nova.network.neutron [-] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 862.152803] env[62460]: DEBUG nova.compute.manager [req-13a7dd38-5dcc-483a-a5f5-3c6de98a23b0 req-89be330b-2a4a-4c97-9600-33d6b5847f12 service nova] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] Received event network-vif-plugged-948f172e-f4a4-484c-b845-ccb61e37c1b2 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 862.153157] env[62460]: DEBUG oslo_concurrency.lockutils [req-13a7dd38-5dcc-483a-a5f5-3c6de98a23b0 req-89be330b-2a4a-4c97-9600-33d6b5847f12 service nova] Acquiring lock "0f5457a0-c918-4407-9ba9-bfeeb949c6a7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.153407] env[62460]: DEBUG oslo_concurrency.lockutils [req-13a7dd38-5dcc-483a-a5f5-3c6de98a23b0 req-89be330b-2a4a-4c97-9600-33d6b5847f12 service nova] Lock "0f5457a0-c918-4407-9ba9-bfeeb949c6a7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.153776] env[62460]: DEBUG oslo_concurrency.lockutils [req-13a7dd38-5dcc-483a-a5f5-3c6de98a23b0 req-89be330b-2a4a-4c97-9600-33d6b5847f12 service nova] Lock "0f5457a0-c918-4407-9ba9-bfeeb949c6a7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.154024] env[62460]: DEBUG nova.compute.manager [req-13a7dd38-5dcc-483a-a5f5-3c6de98a23b0 req-89be330b-2a4a-4c97-9600-33d6b5847f12 service nova] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] No waiting events found dispatching network-vif-plugged-948f172e-f4a4-484c-b845-ccb61e37c1b2 {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 862.154258] env[62460]: WARNING nova.compute.manager [req-13a7dd38-5dcc-483a-a5f5-3c6de98a23b0 req-89be330b-2a4a-4c97-9600-33d6b5847f12 service nova] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] Received unexpected event network-vif-plugged-948f172e-f4a4-484c-b845-ccb61e37c1b2 for instance with vm_state building and task_state spawning. [ 862.154516] env[62460]: DEBUG nova.compute.manager [req-13a7dd38-5dcc-483a-a5f5-3c6de98a23b0 req-89be330b-2a4a-4c97-9600-33d6b5847f12 service nova] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] Received event network-changed-948f172e-f4a4-484c-b845-ccb61e37c1b2 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 862.154777] env[62460]: DEBUG nova.compute.manager [req-13a7dd38-5dcc-483a-a5f5-3c6de98a23b0 req-89be330b-2a4a-4c97-9600-33d6b5847f12 service nova] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] Refreshing instance network info cache due to event network-changed-948f172e-f4a4-484c-b845-ccb61e37c1b2. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 862.155122] env[62460]: DEBUG oslo_concurrency.lockutils [req-13a7dd38-5dcc-483a-a5f5-3c6de98a23b0 req-89be330b-2a4a-4c97-9600-33d6b5847f12 service nova] Acquiring lock "refresh_cache-0f5457a0-c918-4407-9ba9-bfeeb949c6a7" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.155740] env[62460]: DEBUG oslo_concurrency.lockutils [req-13a7dd38-5dcc-483a-a5f5-3c6de98a23b0 req-89be330b-2a4a-4c97-9600-33d6b5847f12 service nova] Acquired lock "refresh_cache-0f5457a0-c918-4407-9ba9-bfeeb949c6a7" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.155740] env[62460]: DEBUG nova.network.neutron [req-13a7dd38-5dcc-483a-a5f5-3c6de98a23b0 req-89be330b-2a4a-4c97-9600-33d6b5847f12 service nova] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] Refreshing network info cache for port 948f172e-f4a4-484c-b845-ccb61e37c1b2 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 862.181885] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313701, 'name': CreateVM_Task, 'duration_secs': 0.409063} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.184998] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 862.189871] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.190024] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.190405] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 862.190743] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d6f84353-f954-40d9-824a-52a6651c8121 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.198788] env[62460]: DEBUG oslo_vmware.api [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Task: {'id': task-1313702, 'name': Rename_Task, 'duration_secs': 0.169764} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.203552] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 862.204038] env[62460]: DEBUG oslo_vmware.api [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Waiting for the task: (returnval){ [ 862.204038] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]526f4bd4-05c6-81bb-5690-2e129751e932" [ 862.204038] env[62460]: _type = "Task" [ 862.204038] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.204795] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1a8ecbe7-6158-4c56-bbea-535dba8a8ea8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.215550] env[62460]: DEBUG oslo_vmware.api [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]526f4bd4-05c6-81bb-5690-2e129751e932, 'name': SearchDatastore_Task, 'duration_secs': 0.014075} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.219383] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.219644] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 862.219887] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.220057] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.220248] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 862.220575] env[62460]: DEBUG oslo_vmware.api [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Waiting for the task: (returnval){ [ 862.220575] env[62460]: value = "task-1313703" [ 862.220575] env[62460]: _type = "Task" [ 862.220575] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.220978] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b78e6add-a8ed-4737-b651-f69e70e1361e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.232157] env[62460]: DEBUG oslo_vmware.api [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Task: {'id': task-1313703, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.233533] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 862.233732] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 862.234477] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6e90f73f-159b-4c16-8e51-933e078a1e0e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.243245] env[62460]: DEBUG oslo_vmware.api [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Waiting for the task: (returnval){ [ 862.243245] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]5212f083-cf5f-02c1-f389-c20b35300033" [ 862.243245] env[62460]: _type = "Task" [ 862.243245] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.252249] env[62460]: DEBUG oslo_vmware.api [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5212f083-cf5f-02c1-f389-c20b35300033, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.277453] env[62460]: DEBUG nova.compute.manager [req-7b8b65c6-bef2-4a05-9b76-88b0310c371f req-23be3ccb-2887-4862-9e9e-fdc8bf214075 service nova] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Received event network-changed-72a87ddf-0585-429a-b9de-d73bcad42cd1 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 862.277594] env[62460]: DEBUG nova.compute.manager [req-7b8b65c6-bef2-4a05-9b76-88b0310c371f req-23be3ccb-2887-4862-9e9e-fdc8bf214075 service nova] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Refreshing instance network info cache due to event network-changed-72a87ddf-0585-429a-b9de-d73bcad42cd1. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 862.278743] env[62460]: DEBUG oslo_concurrency.lockutils [req-7b8b65c6-bef2-4a05-9b76-88b0310c371f req-23be3ccb-2887-4862-9e9e-fdc8bf214075 service nova] Acquiring lock "refresh_cache-896f8ff4-e45f-4403-a727-03ee25e58609" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.278743] env[62460]: DEBUG oslo_concurrency.lockutils [req-7b8b65c6-bef2-4a05-9b76-88b0310c371f req-23be3ccb-2887-4862-9e9e-fdc8bf214075 service nova] Acquired lock "refresh_cache-896f8ff4-e45f-4403-a727-03ee25e58609" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.278743] env[62460]: DEBUG nova.network.neutron [req-7b8b65c6-bef2-4a05-9b76-88b0310c371f req-23be3ccb-2887-4862-9e9e-fdc8bf214075 service nova] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Refreshing network info cache for port 72a87ddf-0585-429a-b9de-d73bcad42cd1 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 862.436007] env[62460]: DEBUG oslo_vmware.api [None req-7ec877fc-e0a0-40d7-aa37-b25eed5a9144 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Task: {'id': task-1313697, 'name': CloneVM_Task} progress is 94%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.437609] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d1c0fc0-1071-45bf-b355-a25d7732eeaf {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.444451] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-939a9afd-5782-45fd-89ad-e6ea936fe49a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.473817] env[62460]: DEBUG nova.network.neutron [-] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 862.475609] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b132982e-3ccb-417d-98f3-7b62d7d038a9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.483654] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7904056-d055-4a5c-8014-783be239355b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.497808] env[62460]: DEBUG nova.compute.provider_tree [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 862.739150] env[62460]: DEBUG oslo_vmware.api [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Task: {'id': task-1313703, 'name': PowerOnVM_Task} progress is 87%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.756631] env[62460]: DEBUG oslo_vmware.api [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5212f083-cf5f-02c1-f389-c20b35300033, 'name': SearchDatastore_Task, 'duration_secs': 0.014534} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.757960] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2fe38908-a74e-4e42-9720-699448eb112f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.765807] env[62460]: DEBUG oslo_vmware.api [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Waiting for the task: (returnval){ [ 862.765807] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]527a387a-491e-42fe-d3f6-c64755db5dc8" [ 862.765807] env[62460]: _type = "Task" [ 862.765807] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.779118] env[62460]: DEBUG oslo_vmware.api [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]527a387a-491e-42fe-d3f6-c64755db5dc8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.938276] env[62460]: DEBUG oslo_vmware.api [None req-7ec877fc-e0a0-40d7-aa37-b25eed5a9144 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Task: {'id': task-1313697, 'name': CloneVM_Task} progress is 95%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.977064] env[62460]: INFO nova.compute.manager [-] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Took 1.03 seconds to deallocate network for instance. [ 862.994317] env[62460]: DEBUG nova.network.neutron [req-13a7dd38-5dcc-483a-a5f5-3c6de98a23b0 req-89be330b-2a4a-4c97-9600-33d6b5847f12 service nova] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] Updated VIF entry in instance network info cache for port 948f172e-f4a4-484c-b845-ccb61e37c1b2. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 862.994750] env[62460]: DEBUG nova.network.neutron [req-13a7dd38-5dcc-483a-a5f5-3c6de98a23b0 req-89be330b-2a4a-4c97-9600-33d6b5847f12 service nova] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] Updating instance_info_cache with network_info: [{"id": "948f172e-f4a4-484c-b845-ccb61e37c1b2", "address": "fa:16:3e:d1:66:a7", "network": {"id": "c2c0d035-a856-4110-b92d-8a630ba71d3c", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1252073445-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b7482bca612d4745a2b540c320c07ac7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4b43a78-f49b-4132-ab2e-6e28769a9498", "external-id": "nsx-vlan-transportzone-737", "segmentation_id": 737, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap948f172e-f4", "ovs_interfaceid": "948f172e-f4a4-484c-b845-ccb61e37c1b2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 863.001135] env[62460]: DEBUG nova.scheduler.client.report [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 863.091191] env[62460]: DEBUG nova.network.neutron [req-7b8b65c6-bef2-4a05-9b76-88b0310c371f req-23be3ccb-2887-4862-9e9e-fdc8bf214075 service nova] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Updated VIF entry in instance network info cache for port 72a87ddf-0585-429a-b9de-d73bcad42cd1. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 863.091588] env[62460]: DEBUG nova.network.neutron [req-7b8b65c6-bef2-4a05-9b76-88b0310c371f req-23be3ccb-2887-4862-9e9e-fdc8bf214075 service nova] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Updating instance_info_cache with network_info: [{"id": "72a87ddf-0585-429a-b9de-d73bcad42cd1", "address": "fa:16:3e:85:2f:50", "network": {"id": "b5cb4ee0-4562-45be-a25d-51c3a5932efd", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-823340443-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.149", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3d2d1c48ec14121a2e8c9b3f800a949", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55c757ac-f8b2-466d-b634-07dbd100b312", "external-id": "nsx-vlan-transportzone-159", "segmentation_id": 159, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap72a87ddf-05", "ovs_interfaceid": "72a87ddf-0585-429a-b9de-d73bcad42cd1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 863.234023] env[62460]: DEBUG oslo_vmware.api [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Task: {'id': task-1313703, 'name': PowerOnVM_Task, 'duration_secs': 0.609581} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.234023] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 863.235864] env[62460]: INFO nova.compute.manager [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Took 5.26 seconds to spawn the instance on the hypervisor. [ 863.235864] env[62460]: DEBUG nova.compute.manager [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 863.235864] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efa33de7-2c6e-45f8-8e0d-e4794a5d60b3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.277434] env[62460]: DEBUG oslo_vmware.api [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]527a387a-491e-42fe-d3f6-c64755db5dc8, 'name': SearchDatastore_Task, 'duration_secs': 0.016184} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.278422] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.278422] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] 0f5457a0-c918-4407-9ba9-bfeeb949c6a7/0f5457a0-c918-4407-9ba9-bfeeb949c6a7.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 863.279450] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-37700773-a4f0-46dd-9659-be6bf6a468ba {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.287389] env[62460]: DEBUG oslo_vmware.api [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Waiting for the task: (returnval){ [ 863.287389] env[62460]: value = "task-1313704" [ 863.287389] env[62460]: _type = "Task" [ 863.287389] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.297080] env[62460]: DEBUG oslo_vmware.api [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Task: {'id': task-1313704, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.441982] env[62460]: DEBUG oslo_vmware.api [None req-7ec877fc-e0a0-40d7-aa37-b25eed5a9144 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Task: {'id': task-1313697, 'name': CloneVM_Task, 'duration_secs': 1.899178} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.442396] env[62460]: INFO nova.virt.vmwareapi.vmops [None req-7ec877fc-e0a0-40d7-aa37-b25eed5a9144 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Created linked-clone VM from snapshot [ 863.443519] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86dd12b7-3f43-4717-85d0-82fd5de3fdca {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.453246] env[62460]: DEBUG nova.virt.vmwareapi.images [None req-7ec877fc-e0a0-40d7-aa37-b25eed5a9144 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Uploading image 63978316-fe83-40b2-9331-e910edd8aa18 {{(pid=62460) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 863.471147] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ec877fc-e0a0-40d7-aa37-b25eed5a9144 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Destroying the VM {{(pid=62460) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 863.471483] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-edf5cb55-7f9f-41e0-b138-617e36a379c1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.479163] env[62460]: DEBUG oslo_vmware.api [None req-7ec877fc-e0a0-40d7-aa37-b25eed5a9144 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Waiting for the task: (returnval){ [ 863.479163] env[62460]: value = "task-1313705" [ 863.479163] env[62460]: _type = "Task" [ 863.479163] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.485260] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a8b2d4df-1735-4f45-bc97-65d17e36a285 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.490864] env[62460]: DEBUG oslo_vmware.api [None req-7ec877fc-e0a0-40d7-aa37-b25eed5a9144 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Task: {'id': task-1313705, 'name': Destroy_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.497570] env[62460]: DEBUG oslo_concurrency.lockutils [req-13a7dd38-5dcc-483a-a5f5-3c6de98a23b0 req-89be330b-2a4a-4c97-9600-33d6b5847f12 service nova] Releasing lock "refresh_cache-0f5457a0-c918-4407-9ba9-bfeeb949c6a7" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.505636] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.372s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.506226] env[62460]: DEBUG nova.compute.manager [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 863.509072] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7eaf8235-f240-4400-9d00-8b18849f1e54 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.219s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.509242] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7eaf8235-f240-4400-9d00-8b18849f1e54 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.511284] env[62460]: DEBUG oslo_concurrency.lockutils [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 19.047s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.511488] env[62460]: DEBUG nova.objects.instance [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62460) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 863.535537] env[62460]: INFO nova.scheduler.client.report [None req-7eaf8235-f240-4400-9d00-8b18849f1e54 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Deleted allocations for instance 1f318a64-2c38-470b-8fae-4ba4543a5681 [ 863.595199] env[62460]: DEBUG oslo_concurrency.lockutils [req-7b8b65c6-bef2-4a05-9b76-88b0310c371f req-23be3ccb-2887-4862-9e9e-fdc8bf214075 service nova] Releasing lock "refresh_cache-896f8ff4-e45f-4403-a727-03ee25e58609" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.633748] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e6aa4918-c28c-4873-a5ac-d199133b8107 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "interface-7e6ff902-4a04-43d5-9014-38c4ec88efc4-bf0f715f-7988-45f9-81b6-c92688ff60c7" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.634037] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e6aa4918-c28c-4873-a5ac-d199133b8107 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "interface-7e6ff902-4a04-43d5-9014-38c4ec88efc4-bf0f715f-7988-45f9-81b6-c92688ff60c7" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.722939] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4db91be9-d7a0-4b00-8292-14baeb7ea3f2 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Acquiring lock "49477d35-92ea-4f9d-8333-fc23144b7dfc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.723345] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4db91be9-d7a0-4b00-8292-14baeb7ea3f2 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Lock "49477d35-92ea-4f9d-8333-fc23144b7dfc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.723593] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4db91be9-d7a0-4b00-8292-14baeb7ea3f2 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Acquiring lock "49477d35-92ea-4f9d-8333-fc23144b7dfc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.723799] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4db91be9-d7a0-4b00-8292-14baeb7ea3f2 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Lock "49477d35-92ea-4f9d-8333-fc23144b7dfc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.723975] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4db91be9-d7a0-4b00-8292-14baeb7ea3f2 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Lock "49477d35-92ea-4f9d-8333-fc23144b7dfc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.726410] env[62460]: INFO nova.compute.manager [None req-4db91be9-d7a0-4b00-8292-14baeb7ea3f2 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] Terminating instance [ 863.728568] env[62460]: DEBUG nova.compute.manager [None req-4db91be9-d7a0-4b00-8292-14baeb7ea3f2 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 863.728843] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-4db91be9-d7a0-4b00-8292-14baeb7ea3f2 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 863.729777] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92d394cf-bdfa-47b1-baaf-f36252179bf3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.738744] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-4db91be9-d7a0-4b00-8292-14baeb7ea3f2 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 863.739132] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9dd3066e-bfd4-47b0-b26f-3b6ebabfc6b3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.746209] env[62460]: DEBUG oslo_vmware.api [None req-4db91be9-d7a0-4b00-8292-14baeb7ea3f2 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Waiting for the task: (returnval){ [ 863.746209] env[62460]: value = "task-1313706" [ 863.746209] env[62460]: _type = "Task" [ 863.746209] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.756635] env[62460]: INFO nova.compute.manager [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Took 29.44 seconds to build instance. [ 863.761640] env[62460]: DEBUG oslo_vmware.api [None req-4db91be9-d7a0-4b00-8292-14baeb7ea3f2 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Task: {'id': task-1313706, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.797567] env[62460]: DEBUG oslo_vmware.api [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Task: {'id': task-1313704, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.990842] env[62460]: DEBUG oslo_vmware.api [None req-7ec877fc-e0a0-40d7-aa37-b25eed5a9144 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Task: {'id': task-1313705, 'name': Destroy_Task} progress is 33%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.016865] env[62460]: DEBUG nova.compute.utils [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 864.021515] env[62460]: DEBUG nova.compute.manager [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 864.022706] env[62460]: DEBUG nova.network.neutron [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 864.043187] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7eaf8235-f240-4400-9d00-8b18849f1e54 tempest-ServersTestManualDisk-604684613 tempest-ServersTestManualDisk-604684613-project-member] Lock "1f318a64-2c38-470b-8fae-4ba4543a5681" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.581s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.089031] env[62460]: DEBUG nova.policy [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd7ac705ce4dc4f42b89cf70ac3316f71', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '335cfe76a8c44686891e17906c96d158', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 864.137090] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e6aa4918-c28c-4873-a5ac-d199133b8107 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "7e6ff902-4a04-43d5-9014-38c4ec88efc4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.137439] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e6aa4918-c28c-4873-a5ac-d199133b8107 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquired lock "7e6ff902-4a04-43d5-9014-38c4ec88efc4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.138755] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd4b6f7d-f0b2-4e68-a318-e839c63ea110 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.167595] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50a8f74b-41cc-4c4e-a62a-14b7cfc5da66 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.200964] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-e6aa4918-c28c-4873-a5ac-d199133b8107 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Reconfiguring VM to detach interface {{(pid=62460) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 864.201318] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c2076b73-99e2-4c85-8e20-5346f13bde65 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.221815] env[62460]: DEBUG oslo_vmware.api [None req-e6aa4918-c28c-4873-a5ac-d199133b8107 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Waiting for the task: (returnval){ [ 864.221815] env[62460]: value = "task-1313707" [ 864.221815] env[62460]: _type = "Task" [ 864.221815] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.233187] env[62460]: DEBUG oslo_vmware.api [None req-e6aa4918-c28c-4873-a5ac-d199133b8107 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313707, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.257819] env[62460]: DEBUG oslo_vmware.api [None req-4db91be9-d7a0-4b00-8292-14baeb7ea3f2 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Task: {'id': task-1313706, 'name': PowerOffVM_Task, 'duration_secs': 0.257536} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.258249] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-4db91be9-d7a0-4b00-8292-14baeb7ea3f2 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 864.258546] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-4db91be9-d7a0-4b00-8292-14baeb7ea3f2 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 864.259180] env[62460]: DEBUG oslo_concurrency.lockutils [None req-43212c99-2cea-458d-9487-17c88670fd98 tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Lock "f1921907-f69c-49a3-896c-a4b12b526ffa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 54.364s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.259505] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a4419608-8445-4e00-bfb7-50f88d64f473 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.299072] env[62460]: DEBUG oslo_vmware.api [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Task: {'id': task-1313704, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.724371} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.299379] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] 0f5457a0-c918-4407-9ba9-bfeeb949c6a7/0f5457a0-c918-4407-9ba9-bfeeb949c6a7.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 864.299625] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 864.299960] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e85c5aac-125f-4498-b195-24b070133e06 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.306733] env[62460]: DEBUG oslo_vmware.api [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Waiting for the task: (returnval){ [ 864.306733] env[62460]: value = "task-1313709" [ 864.306733] env[62460]: _type = "Task" [ 864.306733] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.314866] env[62460]: DEBUG oslo_vmware.api [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Task: {'id': task-1313709, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.325442] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-4db91be9-d7a0-4b00-8292-14baeb7ea3f2 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 864.325731] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-4db91be9-d7a0-4b00-8292-14baeb7ea3f2 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] Deleting contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 864.325964] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-4db91be9-d7a0-4b00-8292-14baeb7ea3f2 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Deleting the datastore file [datastore1] 49477d35-92ea-4f9d-8333-fc23144b7dfc {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 864.326296] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-85adfe23-7070-40b2-bb5e-8db510444d32 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.333245] env[62460]: DEBUG oslo_vmware.api [None req-4db91be9-d7a0-4b00-8292-14baeb7ea3f2 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Waiting for the task: (returnval){ [ 864.333245] env[62460]: value = "task-1313710" [ 864.333245] env[62460]: _type = "Task" [ 864.333245] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.341802] env[62460]: DEBUG oslo_vmware.api [None req-4db91be9-d7a0-4b00-8292-14baeb7ea3f2 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Task: {'id': task-1313710, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.489946] env[62460]: DEBUG oslo_vmware.api [None req-7ec877fc-e0a0-40d7-aa37-b25eed5a9144 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Task: {'id': task-1313705, 'name': Destroy_Task, 'duration_secs': 0.673961} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.490349] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-7ec877fc-e0a0-40d7-aa37-b25eed5a9144 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Destroyed the VM [ 864.490746] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7ec877fc-e0a0-40d7-aa37-b25eed5a9144 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Deleting Snapshot of the VM instance {{(pid=62460) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 864.491048] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-ca7d1a09-bb4e-4de3-a5bd-166e1d2fb0cf {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.497785] env[62460]: DEBUG oslo_vmware.api [None req-7ec877fc-e0a0-40d7-aa37-b25eed5a9144 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Waiting for the task: (returnval){ [ 864.497785] env[62460]: value = "task-1313711" [ 864.497785] env[62460]: _type = "Task" [ 864.497785] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.505833] env[62460]: DEBUG oslo_vmware.api [None req-7ec877fc-e0a0-40d7-aa37-b25eed5a9144 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Task: {'id': task-1313711, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.522661] env[62460]: DEBUG nova.compute.manager [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 864.529123] env[62460]: DEBUG oslo_concurrency.lockutils [None req-65af48e5-af5b-46ab-98f1-d22cfddf39da tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.016s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.529843] env[62460]: DEBUG oslo_concurrency.lockutils [None req-810313b0-6843-4505-a7de-48602dfc2f7b tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.086s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.534570] env[62460]: DEBUG oslo_concurrency.lockutils [None req-810313b0-6843-4505-a7de-48602dfc2f7b tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.534570] env[62460]: DEBUG oslo_concurrency.lockutils [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.737s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.534998] env[62460]: INFO nova.compute.claims [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 864.574958] env[62460]: INFO nova.scheduler.client.report [None req-810313b0-6843-4505-a7de-48602dfc2f7b tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Deleted allocations for instance fde12685-d3b9-46a0-8931-25b904d4f21e [ 864.705137] env[62460]: DEBUG oslo_vmware.rw_handles [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52542c93-0f0c-d878-684e-d5939a14b76c/disk-0.vmdk. {{(pid=62460) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 864.706294] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19318aca-1e34-48ff-82ed-768f9179267c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.709620] env[62460]: DEBUG nova.network.neutron [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Successfully created port: 530faf0d-567e-464f-99de-5f2f1d7afaf2 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 864.716801] env[62460]: DEBUG oslo_vmware.rw_handles [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52542c93-0f0c-d878-684e-d5939a14b76c/disk-0.vmdk is in state: ready. {{(pid=62460) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 864.716998] env[62460]: ERROR oslo_vmware.rw_handles [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52542c93-0f0c-d878-684e-d5939a14b76c/disk-0.vmdk due to incomplete transfer. [ 864.717298] env[62460]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-6079b502-d986-4d1a-8de6-65ed8ea55bae {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.728299] env[62460]: DEBUG oslo_vmware.rw_handles [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52542c93-0f0c-d878-684e-d5939a14b76c/disk-0.vmdk. {{(pid=62460) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 864.728535] env[62460]: DEBUG nova.virt.vmwareapi.images [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Uploaded image 85c60cb8-8b28-497d-afb9-ec82145a4f6e to the Glance image server {{(pid=62460) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 864.731570] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Destroying the VM {{(pid=62460) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 864.732976] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-1486cbdf-0882-4b7f-8f2f-8a7ffbf7c470 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.736879] env[62460]: DEBUG oslo_vmware.api [None req-e6aa4918-c28c-4873-a5ac-d199133b8107 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313707, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.742612] env[62460]: DEBUG oslo_vmware.api [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 864.742612] env[62460]: value = "task-1313712" [ 864.742612] env[62460]: _type = "Task" [ 864.742612] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.752392] env[62460]: DEBUG oslo_vmware.api [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313712, 'name': Destroy_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.818972] env[62460]: DEBUG oslo_vmware.api [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Task: {'id': task-1313709, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079233} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.818972] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 864.820787] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7f48365-2e00-46b0-b544-51716c24b3fc {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.861023] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] Reconfiguring VM instance instance-0000004c to attach disk [datastore2] 0f5457a0-c918-4407-9ba9-bfeeb949c6a7/0f5457a0-c918-4407-9ba9-bfeeb949c6a7.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 864.863928] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-df03e215-e432-460c-9484-fc4814de03db {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.892091] env[62460]: DEBUG oslo_vmware.api [None req-4db91be9-d7a0-4b00-8292-14baeb7ea3f2 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Task: {'id': task-1313710, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.289713} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.892091] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-4db91be9-d7a0-4b00-8292-14baeb7ea3f2 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 864.892091] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-4db91be9-d7a0-4b00-8292-14baeb7ea3f2 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] Deleted contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 864.892091] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-4db91be9-d7a0-4b00-8292-14baeb7ea3f2 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 864.892091] env[62460]: INFO nova.compute.manager [None req-4db91be9-d7a0-4b00-8292-14baeb7ea3f2 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] Took 1.16 seconds to destroy the instance on the hypervisor. [ 864.892335] env[62460]: DEBUG oslo.service.loopingcall [None req-4db91be9-d7a0-4b00-8292-14baeb7ea3f2 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 864.893056] env[62460]: DEBUG oslo_vmware.api [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Waiting for the task: (returnval){ [ 864.893056] env[62460]: value = "task-1313713" [ 864.893056] env[62460]: _type = "Task" [ 864.893056] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.893257] env[62460]: DEBUG nova.compute.manager [-] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 864.893346] env[62460]: DEBUG nova.network.neutron [-] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 864.908558] env[62460]: DEBUG oslo_vmware.api [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Task: {'id': task-1313713, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.008383] env[62460]: DEBUG oslo_vmware.api [None req-7ec877fc-e0a0-40d7-aa37-b25eed5a9144 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Task: {'id': task-1313711, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.082578] env[62460]: DEBUG oslo_concurrency.lockutils [None req-810313b0-6843-4505-a7de-48602dfc2f7b tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Lock "fde12685-d3b9-46a0-8931-25b904d4f21e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.790s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.146298] env[62460]: DEBUG nova.compute.manager [req-9eb6ba5e-cf28-4096-99af-91f192a1855f req-c84ceef2-71aa-42d0-98f1-0ee3645eb454 service nova] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Received event network-changed-dbacff36-64b1-4fc7-b0a0-dc78fcb72ac2 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 865.146993] env[62460]: DEBUG nova.compute.manager [req-9eb6ba5e-cf28-4096-99af-91f192a1855f req-c84ceef2-71aa-42d0-98f1-0ee3645eb454 service nova] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Refreshing instance network info cache due to event network-changed-dbacff36-64b1-4fc7-b0a0-dc78fcb72ac2. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 865.146993] env[62460]: DEBUG oslo_concurrency.lockutils [req-9eb6ba5e-cf28-4096-99af-91f192a1855f req-c84ceef2-71aa-42d0-98f1-0ee3645eb454 service nova] Acquiring lock "refresh_cache-f1921907-f69c-49a3-896c-a4b12b526ffa" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.146993] env[62460]: DEBUG oslo_concurrency.lockutils [req-9eb6ba5e-cf28-4096-99af-91f192a1855f req-c84ceef2-71aa-42d0-98f1-0ee3645eb454 service nova] Acquired lock "refresh_cache-f1921907-f69c-49a3-896c-a4b12b526ffa" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.146993] env[62460]: DEBUG nova.network.neutron [req-9eb6ba5e-cf28-4096-99af-91f192a1855f req-c84ceef2-71aa-42d0-98f1-0ee3645eb454 service nova] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Refreshing network info cache for port dbacff36-64b1-4fc7-b0a0-dc78fcb72ac2 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 865.237796] env[62460]: DEBUG oslo_vmware.api [None req-e6aa4918-c28c-4873-a5ac-d199133b8107 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313707, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.252846] env[62460]: DEBUG oslo_vmware.api [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313712, 'name': Destroy_Task, 'duration_secs': 0.351736} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.253051] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Destroyed the VM [ 865.253300] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Deleting Snapshot of the VM instance {{(pid=62460) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 865.257056] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-c3ca1e33-3feb-4502-8f1c-2d65702ebafa {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.260992] env[62460]: DEBUG oslo_vmware.api [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 865.260992] env[62460]: value = "task-1313714" [ 865.260992] env[62460]: _type = "Task" [ 865.260992] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.269342] env[62460]: DEBUG oslo_vmware.api [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313714, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.412514] env[62460]: DEBUG oslo_vmware.api [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Task: {'id': task-1313713, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.508746] env[62460]: DEBUG oslo_vmware.api [None req-7ec877fc-e0a0-40d7-aa37-b25eed5a9144 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Task: {'id': task-1313711, 'name': RemoveSnapshot_Task, 'duration_secs': 0.744927} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.509117] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7ec877fc-e0a0-40d7-aa37-b25eed5a9144 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Deleted Snapshot of the VM instance {{(pid=62460) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 865.534047] env[62460]: DEBUG nova.compute.manager [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 865.586416] env[62460]: DEBUG nova.virt.hardware [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 865.586416] env[62460]: DEBUG nova.virt.hardware [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 865.586416] env[62460]: DEBUG nova.virt.hardware [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 865.586416] env[62460]: DEBUG nova.virt.hardware [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 865.586416] env[62460]: DEBUG nova.virt.hardware [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 865.586416] env[62460]: DEBUG nova.virt.hardware [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 865.586416] env[62460]: DEBUG nova.virt.hardware [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 865.586416] env[62460]: DEBUG nova.virt.hardware [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 865.586972] env[62460]: DEBUG nova.virt.hardware [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 865.586972] env[62460]: DEBUG nova.virt.hardware [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 865.587481] env[62460]: DEBUG nova.virt.hardware [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 865.588820] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bb5bd57-f1df-473a-af3f-23650f834563 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.601711] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86fe0976-4a78-4b84-9b7c-86c630f15d13 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.710405] env[62460]: DEBUG nova.network.neutron [-] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.739135] env[62460]: DEBUG oslo_vmware.api [None req-e6aa4918-c28c-4873-a5ac-d199133b8107 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313707, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.773886] env[62460]: DEBUG oslo_vmware.api [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313714, 'name': RemoveSnapshot_Task, 'duration_secs': 0.433612} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.776030] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Deleted Snapshot of the VM instance {{(pid=62460) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 865.776380] env[62460]: DEBUG nova.compute.manager [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 865.777550] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b421a8a-e7ab-4e36-a4ce-523e6bc240a8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.828386] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0d933b5-d19f-4e2a-a2e3-39cdce2d410f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.838307] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bd4f528-934a-46ea-84dc-c6816c21122c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.881497] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c402b2c-53e5-4426-b749-943236501d2c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.889347] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19674349-eb9f-4ca7-afa5-94f92fdb0b9a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.908290] env[62460]: DEBUG nova.compute.provider_tree [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 865.918700] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b696d9d5-b153-4d81-a79f-4ae255c22926 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Acquiring lock "3ada3516-3147-4566-a46a-1cb29cf880d0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.919060] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b696d9d5-b153-4d81-a79f-4ae255c22926 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Lock "3ada3516-3147-4566-a46a-1cb29cf880d0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.919334] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b696d9d5-b153-4d81-a79f-4ae255c22926 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Acquiring lock "3ada3516-3147-4566-a46a-1cb29cf880d0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.919602] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b696d9d5-b153-4d81-a79f-4ae255c22926 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Lock "3ada3516-3147-4566-a46a-1cb29cf880d0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.920121] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b696d9d5-b153-4d81-a79f-4ae255c22926 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Lock "3ada3516-3147-4566-a46a-1cb29cf880d0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.922619] env[62460]: DEBUG oslo_vmware.api [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Task: {'id': task-1313713, 'name': ReconfigVM_Task, 'duration_secs': 0.657632} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.923547] env[62460]: INFO nova.compute.manager [None req-b696d9d5-b153-4d81-a79f-4ae255c22926 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] Terminating instance [ 865.924941] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] Reconfigured VM instance instance-0000004c to attach disk [datastore2] 0f5457a0-c918-4407-9ba9-bfeeb949c6a7/0f5457a0-c918-4407-9ba9-bfeeb949c6a7.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 865.928950] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-28450865-eb42-4f16-9cf3-b91dd8986083 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.928950] env[62460]: DEBUG nova.compute.manager [None req-b696d9d5-b153-4d81-a79f-4ae255c22926 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 865.928950] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-b696d9d5-b153-4d81-a79f-4ae255c22926 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 865.929294] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79bfcc1f-e29b-41c9-89ec-90f92152baa4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.934800] env[62460]: DEBUG nova.network.neutron [req-9eb6ba5e-cf28-4096-99af-91f192a1855f req-c84ceef2-71aa-42d0-98f1-0ee3645eb454 service nova] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Updated VIF entry in instance network info cache for port dbacff36-64b1-4fc7-b0a0-dc78fcb72ac2. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 865.934800] env[62460]: DEBUG nova.network.neutron [req-9eb6ba5e-cf28-4096-99af-91f192a1855f req-c84ceef2-71aa-42d0-98f1-0ee3645eb454 service nova] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Updating instance_info_cache with network_info: [{"id": "dbacff36-64b1-4fc7-b0a0-dc78fcb72ac2", "address": "fa:16:3e:5e:f0:fa", "network": {"id": "828c25a7-673d-4b6c-ad50-7317affa9ded", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-80559771-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.241", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a10038b038254e5d9543018e39a56992", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e39ca24f-7890-4cdf-8dab-ecab218bb063", "external-id": "nsx-vlan-transportzone-890", "segmentation_id": 890, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdbacff36-64", "ovs_interfaceid": "dbacff36-64b1-4fc7-b0a0-dc78fcb72ac2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.939436] env[62460]: DEBUG oslo_vmware.api [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Waiting for the task: (returnval){ [ 865.939436] env[62460]: value = "task-1313715" [ 865.939436] env[62460]: _type = "Task" [ 865.939436] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.941686] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-b696d9d5-b153-4d81-a79f-4ae255c22926 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 865.945124] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f2b38c32-08b3-45c9-8982-ce1b597a7b4a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.953940] env[62460]: DEBUG oslo_vmware.api [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Task: {'id': task-1313715, 'name': Rename_Task} progress is 6%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.955640] env[62460]: DEBUG oslo_vmware.api [None req-b696d9d5-b153-4d81-a79f-4ae255c22926 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Waiting for the task: (returnval){ [ 865.955640] env[62460]: value = "task-1313716" [ 865.955640] env[62460]: _type = "Task" [ 865.955640] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.965647] env[62460]: DEBUG oslo_vmware.api [None req-b696d9d5-b153-4d81-a79f-4ae255c22926 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313716, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.015484] env[62460]: WARNING nova.compute.manager [None req-7ec877fc-e0a0-40d7-aa37-b25eed5a9144 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Image not found during snapshot: nova.exception.ImageNotFound: Image 63978316-fe83-40b2-9331-e910edd8aa18 could not be found. [ 866.214771] env[62460]: INFO nova.compute.manager [-] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] Took 1.32 seconds to deallocate network for instance. [ 866.240564] env[62460]: DEBUG oslo_vmware.api [None req-e6aa4918-c28c-4873-a5ac-d199133b8107 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313707, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.289917] env[62460]: INFO nova.compute.manager [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Shelve offloading [ 866.291716] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 866.292372] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e1ebfff3-852d-44be-bc5f-74468acfa4f0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.299012] env[62460]: DEBUG oslo_vmware.api [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 866.299012] env[62460]: value = "task-1313717" [ 866.299012] env[62460]: _type = "Task" [ 866.299012] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.309262] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] VM already powered off {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 866.309504] env[62460]: DEBUG nova.compute.manager [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 866.310305] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ec9cad0-05ef-4be2-8e63-dee776ee75d0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.316248] env[62460]: DEBUG oslo_concurrency.lockutils [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "refresh_cache-d7d180f7-e1a8-46c8-ba8e-ca50dac474cc" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 866.316458] env[62460]: DEBUG oslo_concurrency.lockutils [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquired lock "refresh_cache-d7d180f7-e1a8-46c8-ba8e-ca50dac474cc" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.316655] env[62460]: DEBUG nova.network.neutron [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 866.414792] env[62460]: DEBUG nova.scheduler.client.report [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 866.437136] env[62460]: DEBUG oslo_concurrency.lockutils [req-9eb6ba5e-cf28-4096-99af-91f192a1855f req-c84ceef2-71aa-42d0-98f1-0ee3645eb454 service nova] Releasing lock "refresh_cache-f1921907-f69c-49a3-896c-a4b12b526ffa" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.454187] env[62460]: DEBUG oslo_vmware.api [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Task: {'id': task-1313715, 'name': Rename_Task, 'duration_secs': 0.303199} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.454549] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 866.455041] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e94ed709-da64-483b-8a12-86dd1cffc6cd {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.470733] env[62460]: DEBUG oslo_vmware.api [None req-b696d9d5-b153-4d81-a79f-4ae255c22926 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313716, 'name': PowerOffVM_Task, 'duration_secs': 0.384392} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.471613] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-b696d9d5-b153-4d81-a79f-4ae255c22926 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 866.471764] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-b696d9d5-b153-4d81-a79f-4ae255c22926 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 866.472090] env[62460]: DEBUG oslo_vmware.api [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Waiting for the task: (returnval){ [ 866.472090] env[62460]: value = "task-1313718" [ 866.472090] env[62460]: _type = "Task" [ 866.472090] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.472311] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cd4e8869-236d-4488-9a16-440a62e86967 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.481501] env[62460]: DEBUG oslo_vmware.api [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Task: {'id': task-1313718, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.547230] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-b696d9d5-b153-4d81-a79f-4ae255c22926 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 866.548505] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-b696d9d5-b153-4d81-a79f-4ae255c22926 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] Deleting contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 866.548505] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-b696d9d5-b153-4d81-a79f-4ae255c22926 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Deleting the datastore file [datastore2] 3ada3516-3147-4566-a46a-1cb29cf880d0 {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 866.548505] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-537e511e-6e41-45c0-95c6-994e78a041a0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.557855] env[62460]: DEBUG oslo_vmware.api [None req-b696d9d5-b153-4d81-a79f-4ae255c22926 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Waiting for the task: (returnval){ [ 866.557855] env[62460]: value = "task-1313720" [ 866.557855] env[62460]: _type = "Task" [ 866.557855] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.568027] env[62460]: DEBUG oslo_vmware.api [None req-b696d9d5-b153-4d81-a79f-4ae255c22926 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313720, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.722814] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4db91be9-d7a0-4b00-8292-14baeb7ea3f2 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.741021] env[62460]: DEBUG oslo_vmware.api [None req-e6aa4918-c28c-4873-a5ac-d199133b8107 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313707, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.894081] env[62460]: DEBUG oslo_concurrency.lockutils [None req-36066889-905f-461f-99f9-b5b6a17671e2 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Acquiring lock "a02b4be1-91b0-4254-8d60-654885e24f6b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.894934] env[62460]: DEBUG oslo_concurrency.lockutils [None req-36066889-905f-461f-99f9-b5b6a17671e2 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Lock "a02b4be1-91b0-4254-8d60-654885e24f6b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.895298] env[62460]: DEBUG oslo_concurrency.lockutils [None req-36066889-905f-461f-99f9-b5b6a17671e2 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Acquiring lock "a02b4be1-91b0-4254-8d60-654885e24f6b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.895415] env[62460]: DEBUG oslo_concurrency.lockutils [None req-36066889-905f-461f-99f9-b5b6a17671e2 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Lock "a02b4be1-91b0-4254-8d60-654885e24f6b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.895877] env[62460]: DEBUG oslo_concurrency.lockutils [None req-36066889-905f-461f-99f9-b5b6a17671e2 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Lock "a02b4be1-91b0-4254-8d60-654885e24f6b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.898221] env[62460]: INFO nova.compute.manager [None req-36066889-905f-461f-99f9-b5b6a17671e2 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Terminating instance [ 866.900364] env[62460]: DEBUG nova.compute.manager [None req-36066889-905f-461f-99f9-b5b6a17671e2 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 866.900697] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-36066889-905f-461f-99f9-b5b6a17671e2 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 866.901706] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f47ce907-0200-44eb-8d0b-467d58078def {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.909596] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-36066889-905f-461f-99f9-b5b6a17671e2 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 866.914899] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8dd595d4-4f1c-4207-a50e-1a1b607c3ede {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.920165] env[62460]: DEBUG oslo_vmware.api [None req-36066889-905f-461f-99f9-b5b6a17671e2 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Waiting for the task: (returnval){ [ 866.920165] env[62460]: value = "task-1313721" [ 866.920165] env[62460]: _type = "Task" [ 866.920165] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.920930] env[62460]: DEBUG oslo_concurrency.lockutils [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.388s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.921558] env[62460]: DEBUG nova.compute.manager [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 866.927738] env[62460]: DEBUG oslo_concurrency.lockutils [None req-759824f9-86b1-41cc-87e9-b267e31f479b tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.053s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.928033] env[62460]: DEBUG oslo_concurrency.lockutils [None req-759824f9-86b1-41cc-87e9-b267e31f479b tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.929915] env[62460]: DEBUG oslo_concurrency.lockutils [None req-af702ee7-e128-4206-aea5-9106143dcc05 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.805s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.930077] env[62460]: DEBUG oslo_concurrency.lockutils [None req-af702ee7-e128-4206-aea5-9106143dcc05 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.932124] env[62460]: DEBUG oslo_concurrency.lockutils [None req-53ff2f89-d24e-47c8-a414-09b17968e3cd tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.696s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.932198] env[62460]: DEBUG oslo_concurrency.lockutils [None req-53ff2f89-d24e-47c8-a414-09b17968e3cd tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.933779] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 9.483s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.933978] env[62460]: DEBUG nova.objects.instance [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62460) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 866.944427] env[62460]: DEBUG nova.compute.manager [req-32281eb3-970c-40fd-afe6-89f3418e79ca req-dded8dc7-89cf-4c58-97a1-b3e40783b3c4 service nova] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Received event network-vif-plugged-530faf0d-567e-464f-99de-5f2f1d7afaf2 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 866.944427] env[62460]: DEBUG oslo_concurrency.lockutils [req-32281eb3-970c-40fd-afe6-89f3418e79ca req-dded8dc7-89cf-4c58-97a1-b3e40783b3c4 service nova] Acquiring lock "3e10af70-db52-4d4a-bb92-821a05dcbab6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.944427] env[62460]: DEBUG oslo_concurrency.lockutils [req-32281eb3-970c-40fd-afe6-89f3418e79ca req-dded8dc7-89cf-4c58-97a1-b3e40783b3c4 service nova] Lock "3e10af70-db52-4d4a-bb92-821a05dcbab6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.944427] env[62460]: DEBUG oslo_concurrency.lockutils [req-32281eb3-970c-40fd-afe6-89f3418e79ca req-dded8dc7-89cf-4c58-97a1-b3e40783b3c4 service nova] Lock "3e10af70-db52-4d4a-bb92-821a05dcbab6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.944427] env[62460]: DEBUG nova.compute.manager [req-32281eb3-970c-40fd-afe6-89f3418e79ca req-dded8dc7-89cf-4c58-97a1-b3e40783b3c4 service nova] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] No waiting events found dispatching network-vif-plugged-530faf0d-567e-464f-99de-5f2f1d7afaf2 {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 866.944427] env[62460]: WARNING nova.compute.manager [req-32281eb3-970c-40fd-afe6-89f3418e79ca req-dded8dc7-89cf-4c58-97a1-b3e40783b3c4 service nova] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Received unexpected event network-vif-plugged-530faf0d-567e-464f-99de-5f2f1d7afaf2 for instance with vm_state building and task_state spawning. [ 866.953349] env[62460]: DEBUG oslo_vmware.api [None req-36066889-905f-461f-99f9-b5b6a17671e2 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Task: {'id': task-1313721, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.986325] env[62460]: DEBUG oslo_vmware.api [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Task: {'id': task-1313718, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.987298] env[62460]: INFO nova.scheduler.client.report [None req-af702ee7-e128-4206-aea5-9106143dcc05 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Deleted allocations for instance 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b [ 866.988928] env[62460]: INFO nova.scheduler.client.report [None req-759824f9-86b1-41cc-87e9-b267e31f479b tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Deleted allocations for instance b76f37a0-91d0-4a01-9d95-9c6586081175 [ 867.011258] env[62460]: INFO nova.scheduler.client.report [None req-53ff2f89-d24e-47c8-a414-09b17968e3cd tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Deleted allocations for instance 0269dc64-d2b1-43c5-bdf7-11d97e534819 [ 867.070308] env[62460]: DEBUG oslo_vmware.api [None req-b696d9d5-b153-4d81-a79f-4ae255c22926 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313720, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.150526} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.070702] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-b696d9d5-b153-4d81-a79f-4ae255c22926 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 867.070917] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-b696d9d5-b153-4d81-a79f-4ae255c22926 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] Deleted contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 867.071188] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-b696d9d5-b153-4d81-a79f-4ae255c22926 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 867.071388] env[62460]: INFO nova.compute.manager [None req-b696d9d5-b153-4d81-a79f-4ae255c22926 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] Took 1.14 seconds to destroy the instance on the hypervisor. [ 867.071717] env[62460]: DEBUG oslo.service.loopingcall [None req-b696d9d5-b153-4d81-a79f-4ae255c22926 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 867.071904] env[62460]: DEBUG nova.compute.manager [-] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 867.072016] env[62460]: DEBUG nova.network.neutron [-] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 867.082523] env[62460]: DEBUG nova.network.neutron [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Successfully updated port: 530faf0d-567e-464f-99de-5f2f1d7afaf2 {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 867.169038] env[62460]: DEBUG nova.network.neutron [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Updating instance_info_cache with network_info: [{"id": "5606455d-d407-41ff-8232-841d166aa4b1", "address": "fa:16:3e:6d:35:05", "network": {"id": "de789887-9cf9-4614-a7e1-dab430aaabaf", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-656400207-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0da056d93bdf40c39d6e82e457727ff6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5116f690-f825-4fee-8a47-42b073e716c5", "external-id": "nsx-vlan-transportzone-692", "segmentation_id": 692, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5606455d-d4", "ovs_interfaceid": "5606455d-d407-41ff-8232-841d166aa4b1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.183965] env[62460]: DEBUG nova.compute.manager [req-b7818f8b-3fde-44a8-ad5f-667a869b226b req-a3379c47-d8e3-4f8e-99aa-4d6467300437 service nova] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] Received event network-vif-deleted-00526c55-f6d3-48d6-ba66-a1f821f2b03b {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 867.241860] env[62460]: DEBUG oslo_vmware.api [None req-e6aa4918-c28c-4873-a5ac-d199133b8107 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313707, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.432869] env[62460]: DEBUG oslo_vmware.api [None req-36066889-905f-461f-99f9-b5b6a17671e2 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Task: {'id': task-1313721, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.438508] env[62460]: DEBUG nova.compute.utils [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 867.439997] env[62460]: DEBUG nova.compute.manager [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 867.440228] env[62460]: DEBUG nova.network.neutron [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 867.485059] env[62460]: DEBUG oslo_vmware.api [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Task: {'id': task-1313718, 'name': PowerOnVM_Task, 'duration_secs': 0.905398} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.486649] env[62460]: DEBUG nova.policy [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e531b7c3e7544e53b126d1336ac44c99', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bfa5a65b0f614e769de5b3aa77bf869e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 867.488280] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 867.488512] env[62460]: INFO nova.compute.manager [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] Took 8.33 seconds to spawn the instance on the hypervisor. [ 867.488699] env[62460]: DEBUG nova.compute.manager [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 867.489811] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-285f1188-74bc-40ce-b342-9f8c7db171f2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.503270] env[62460]: DEBUG oslo_concurrency.lockutils [None req-759824f9-86b1-41cc-87e9-b267e31f479b tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "b76f37a0-91d0-4a01-9d95-9c6586081175" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.836s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.504482] env[62460]: DEBUG oslo_concurrency.lockutils [None req-af702ee7-e128-4206-aea5-9106143dcc05 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Lock "9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.968s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.519034] env[62460]: DEBUG oslo_concurrency.lockutils [None req-53ff2f89-d24e-47c8-a414-09b17968e3cd tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Lock "0269dc64-d2b1-43c5-bdf7-11d97e534819" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.675s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.585591] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Acquiring lock "refresh_cache-3e10af70-db52-4d4a-bb92-821a05dcbab6" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.586816] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Acquired lock "refresh_cache-3e10af70-db52-4d4a-bb92-821a05dcbab6" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.586816] env[62460]: DEBUG nova.network.neutron [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 867.671775] env[62460]: DEBUG oslo_concurrency.lockutils [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Releasing lock "refresh_cache-d7d180f7-e1a8-46c8-ba8e-ca50dac474cc" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.738118] env[62460]: DEBUG oslo_vmware.api [None req-e6aa4918-c28c-4873-a5ac-d199133b8107 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313707, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.856733] env[62460]: DEBUG nova.network.neutron [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Successfully created port: f5fba0f8-a3a5-4bcf-856c-52e2f4313154 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 867.885435] env[62460]: DEBUG nova.network.neutron [-] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.930563] env[62460]: DEBUG oslo_vmware.api [None req-36066889-905f-461f-99f9-b5b6a17671e2 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Task: {'id': task-1313721, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.942985] env[62460]: DEBUG nova.compute.manager [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 867.949262] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3891b63c-5086-4f3a-9d56-cd1b3d4802c4 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.015s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.950389] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5770ec75-799a-4db6-97a5-b98712f6f11d tempest-ServersListShow296Test-180931441 tempest-ServersListShow296Test-180931441-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.473s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.955065] env[62460]: INFO nova.compute.claims [None req-5770ec75-799a-4db6-97a5-b98712f6f11d tempest-ServersListShow296Test-180931441 tempest-ServersListShow296Test-180931441-project-member] [instance: 120b7508-dc38-4659-92cb-8b3e813dae2a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 867.991514] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 867.992530] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea4adb61-933e-4929-ba18-8472fe400a12 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.002236] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 868.002469] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-868d2698-a6f0-4975-984b-da13c7acb880 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.017875] env[62460]: INFO nova.compute.manager [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] Took 31.14 seconds to build instance. [ 868.099068] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 868.099329] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Deleting contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 868.099523] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Deleting the datastore file [datastore2] d7d180f7-e1a8-46c8-ba8e-ca50dac474cc {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 868.099796] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2c11a0d0-5c09-4118-975f-5135752cc27b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.107686] env[62460]: DEBUG oslo_vmware.api [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 868.107686] env[62460]: value = "task-1313723" [ 868.107686] env[62460]: _type = "Task" [ 868.107686] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.118343] env[62460]: DEBUG oslo_vmware.api [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313723, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.147190] env[62460]: DEBUG nova.network.neutron [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 868.241144] env[62460]: DEBUG oslo_vmware.api [None req-e6aa4918-c28c-4873-a5ac-d199133b8107 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313707, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.389151] env[62460]: INFO nova.compute.manager [-] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] Took 1.32 seconds to deallocate network for instance. [ 868.434052] env[62460]: DEBUG oslo_vmware.api [None req-36066889-905f-461f-99f9-b5b6a17671e2 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Task: {'id': task-1313721, 'name': PowerOffVM_Task, 'duration_secs': 1.20196} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.434052] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-36066889-905f-461f-99f9-b5b6a17671e2 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 868.434052] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-36066889-905f-461f-99f9-b5b6a17671e2 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 868.434052] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-88222b6b-39db-4ff9-8e72-4d1ba6c64f39 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.453327] env[62460]: DEBUG nova.network.neutron [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Updating instance_info_cache with network_info: [{"id": "530faf0d-567e-464f-99de-5f2f1d7afaf2", "address": "fa:16:3e:7d:d3:e1", "network": {"id": "51277201-fa18-437f-95a2-e42936d56791", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1816940505-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "335cfe76a8c44686891e17906c96d158", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "84aee122-f630-43c5-9cc1-3a38d3819c82", "external-id": "nsx-vlan-transportzone-816", "segmentation_id": 816, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap530faf0d-56", "ovs_interfaceid": "530faf0d-567e-464f-99de-5f2f1d7afaf2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.460483] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5770ec75-799a-4db6-97a5-b98712f6f11d tempest-ServersListShow296Test-180931441 tempest-ServersListShow296Test-180931441-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 0.509s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.460483] env[62460]: DEBUG nova.compute.utils [None req-5770ec75-799a-4db6-97a5-b98712f6f11d tempest-ServersListShow296Test-180931441 tempest-ServersListShow296Test-180931441-project-member] [instance: 120b7508-dc38-4659-92cb-8b3e813dae2a] Instance 120b7508-dc38-4659-92cb-8b3e813dae2a could not be found. {{(pid=62460) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 868.460892] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a0a3bf88-42e5-4bc9-9abb-b76cbcd709f3 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.072s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.461136] env[62460]: DEBUG nova.objects.instance [None req-a0a3bf88-42e5-4bc9-9abb-b76cbcd709f3 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lazy-loading 'resources' on Instance uuid 16d40829-ba6e-4193-98b5-fff94c066bcb {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 868.462837] env[62460]: DEBUG nova.compute.manager [None req-5770ec75-799a-4db6-97a5-b98712f6f11d tempest-ServersListShow296Test-180931441 tempest-ServersListShow296Test-180931441-project-member] [instance: 120b7508-dc38-4659-92cb-8b3e813dae2a] Instance disappeared during build. {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2513}} [ 868.463062] env[62460]: DEBUG nova.compute.manager [None req-5770ec75-799a-4db6-97a5-b98712f6f11d tempest-ServersListShow296Test-180931441 tempest-ServersListShow296Test-180931441-project-member] [instance: 120b7508-dc38-4659-92cb-8b3e813dae2a] Unplugging VIFs for instance {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 868.463259] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5770ec75-799a-4db6-97a5-b98712f6f11d tempest-ServersListShow296Test-180931441 tempest-ServersListShow296Test-180931441-project-member] Acquiring lock "refresh_cache-120b7508-dc38-4659-92cb-8b3e813dae2a" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.463416] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5770ec75-799a-4db6-97a5-b98712f6f11d tempest-ServersListShow296Test-180931441 tempest-ServersListShow296Test-180931441-project-member] Acquired lock "refresh_cache-120b7508-dc38-4659-92cb-8b3e813dae2a" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.466990] env[62460]: DEBUG nova.network.neutron [None req-5770ec75-799a-4db6-97a5-b98712f6f11d tempest-ServersListShow296Test-180931441 tempest-ServersListShow296Test-180931441-project-member] [instance: 120b7508-dc38-4659-92cb-8b3e813dae2a] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 868.520137] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a548bc15-c883-4a9a-a881-3cd31aecdd32 tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Lock "0f5457a0-c918-4407-9ba9-bfeeb949c6a7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 50.871s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.620123] env[62460]: DEBUG oslo_vmware.api [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313723, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.398745} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.621169] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 868.621576] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Deleted contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 868.622771] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 868.658276] env[62460]: INFO nova.scheduler.client.report [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Deleted allocations for instance d7d180f7-e1a8-46c8-ba8e-ca50dac474cc [ 868.743658] env[62460]: DEBUG oslo_vmware.api [None req-e6aa4918-c28c-4873-a5ac-d199133b8107 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313707, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.898151] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b696d9d5-b153-4d81-a79f-4ae255c22926 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.959823] env[62460]: DEBUG nova.compute.manager [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 868.962521] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Releasing lock "refresh_cache-3e10af70-db52-4d4a-bb92-821a05dcbab6" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.962521] env[62460]: DEBUG nova.compute.manager [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Instance network_info: |[{"id": "530faf0d-567e-464f-99de-5f2f1d7afaf2", "address": "fa:16:3e:7d:d3:e1", "network": {"id": "51277201-fa18-437f-95a2-e42936d56791", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1816940505-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "335cfe76a8c44686891e17906c96d158", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "84aee122-f630-43c5-9cc1-3a38d3819c82", "external-id": "nsx-vlan-transportzone-816", "segmentation_id": 816, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap530faf0d-56", "ovs_interfaceid": "530faf0d-567e-464f-99de-5f2f1d7afaf2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 868.962912] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7d:d3:e1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '84aee122-f630-43c5-9cc1-3a38d3819c82', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '530faf0d-567e-464f-99de-5f2f1d7afaf2', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 868.971779] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Creating folder: Project (335cfe76a8c44686891e17906c96d158). Parent ref: group-v281134. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 868.971779] env[62460]: DEBUG nova.compute.utils [None req-5770ec75-799a-4db6-97a5-b98712f6f11d tempest-ServersListShow296Test-180931441 tempest-ServersListShow296Test-180931441-project-member] [instance: 120b7508-dc38-4659-92cb-8b3e813dae2a] Can not refresh info_cache because instance was not found {{(pid=62460) refresh_info_cache_for_instance /opt/stack/nova/nova/compute/utils.py:1024}} [ 868.974732] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3c9dd034-26d3-4cb3-be65-97445eb62226 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.985575] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Created folder: Project (335cfe76a8c44686891e17906c96d158) in parent group-v281134. [ 868.987068] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Creating folder: Instances. Parent ref: group-v281219. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 868.987068] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1bf49777-5022-446a-91c6-06a214eda9af {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.990031] env[62460]: DEBUG nova.virt.hardware [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 868.990031] env[62460]: DEBUG nova.virt.hardware [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 868.990031] env[62460]: DEBUG nova.virt.hardware [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 868.990031] env[62460]: DEBUG nova.virt.hardware [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 868.990230] env[62460]: DEBUG nova.virt.hardware [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 868.990279] env[62460]: DEBUG nova.virt.hardware [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 868.990755] env[62460]: DEBUG nova.virt.hardware [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 868.990755] env[62460]: DEBUG nova.virt.hardware [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 868.991292] env[62460]: DEBUG nova.virt.hardware [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 868.991292] env[62460]: DEBUG nova.virt.hardware [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 868.991292] env[62460]: DEBUG nova.virt.hardware [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 868.991965] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74c81d0a-c1bf-496e-b2c4-e18bcda4b573 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.002237] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e4f90ab-7a66-4c3f-8f5b-ac829d8143f6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.006385] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Created folder: Instances in parent group-v281219. [ 869.006640] env[62460]: DEBUG oslo.service.loopingcall [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 869.007188] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 869.007431] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7301ad37-3fdd-4d90-ba01-3696e5c9e449 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.024157] env[62460]: DEBUG nova.network.neutron [None req-5770ec75-799a-4db6-97a5-b98712f6f11d tempest-ServersListShow296Test-180931441 tempest-ServersListShow296Test-180931441-project-member] [instance: 120b7508-dc38-4659-92cb-8b3e813dae2a] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 869.039544] env[62460]: DEBUG nova.compute.manager [req-5f0c99a6-c84f-4747-88df-c8e6086001bc req-8cba1c45-a728-4a5c-95a3-df71ac96d327 service nova] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Received event network-changed-530faf0d-567e-464f-99de-5f2f1d7afaf2 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 869.039544] env[62460]: DEBUG nova.compute.manager [req-5f0c99a6-c84f-4747-88df-c8e6086001bc req-8cba1c45-a728-4a5c-95a3-df71ac96d327 service nova] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Refreshing instance network info cache due to event network-changed-530faf0d-567e-464f-99de-5f2f1d7afaf2. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 869.039544] env[62460]: DEBUG oslo_concurrency.lockutils [req-5f0c99a6-c84f-4747-88df-c8e6086001bc req-8cba1c45-a728-4a5c-95a3-df71ac96d327 service nova] Acquiring lock "refresh_cache-3e10af70-db52-4d4a-bb92-821a05dcbab6" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 869.039544] env[62460]: DEBUG oslo_concurrency.lockutils [req-5f0c99a6-c84f-4747-88df-c8e6086001bc req-8cba1c45-a728-4a5c-95a3-df71ac96d327 service nova] Acquired lock "refresh_cache-3e10af70-db52-4d4a-bb92-821a05dcbab6" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.039544] env[62460]: DEBUG nova.network.neutron [req-5f0c99a6-c84f-4747-88df-c8e6086001bc req-8cba1c45-a728-4a5c-95a3-df71ac96d327 service nova] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Refreshing network info cache for port 530faf0d-567e-464f-99de-5f2f1d7afaf2 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 869.040108] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 869.040108] env[62460]: value = "task-1313727" [ 869.040108] env[62460]: _type = "Task" [ 869.040108] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.053647] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313727, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.165899] env[62460]: DEBUG oslo_concurrency.lockutils [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.176021] env[62460]: DEBUG nova.network.neutron [None req-5770ec75-799a-4db6-97a5-b98712f6f11d tempest-ServersListShow296Test-180931441 tempest-ServersListShow296Test-180931441-project-member] [instance: 120b7508-dc38-4659-92cb-8b3e813dae2a] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.243076] env[62460]: DEBUG oslo_vmware.api [None req-e6aa4918-c28c-4873-a5ac-d199133b8107 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313707, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.248262] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b417a114-6bca-4bbd-a3d9-76022a14d1b1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.252682] env[62460]: DEBUG nova.compute.manager [req-09774240-dd45-4916-9b9e-e6f16f650aa5 req-4ba10221-421f-4518-b94f-befebf4a06ac service nova] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Received event network-vif-unplugged-5606455d-d407-41ff-8232-841d166aa4b1 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 869.252852] env[62460]: DEBUG oslo_concurrency.lockutils [req-09774240-dd45-4916-9b9e-e6f16f650aa5 req-4ba10221-421f-4518-b94f-befebf4a06ac service nova] Acquiring lock "d7d180f7-e1a8-46c8-ba8e-ca50dac474cc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.254016] env[62460]: DEBUG oslo_concurrency.lockutils [req-09774240-dd45-4916-9b9e-e6f16f650aa5 req-4ba10221-421f-4518-b94f-befebf4a06ac service nova] Lock "d7d180f7-e1a8-46c8-ba8e-ca50dac474cc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.254016] env[62460]: DEBUG oslo_concurrency.lockutils [req-09774240-dd45-4916-9b9e-e6f16f650aa5 req-4ba10221-421f-4518-b94f-befebf4a06ac service nova] Lock "d7d180f7-e1a8-46c8-ba8e-ca50dac474cc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.254016] env[62460]: DEBUG nova.compute.manager [req-09774240-dd45-4916-9b9e-e6f16f650aa5 req-4ba10221-421f-4518-b94f-befebf4a06ac service nova] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] No waiting events found dispatching network-vif-unplugged-5606455d-d407-41ff-8232-841d166aa4b1 {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 869.254016] env[62460]: WARNING nova.compute.manager [req-09774240-dd45-4916-9b9e-e6f16f650aa5 req-4ba10221-421f-4518-b94f-befebf4a06ac service nova] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Received unexpected event network-vif-unplugged-5606455d-d407-41ff-8232-841d166aa4b1 for instance with vm_state shelved_offloaded and task_state None. [ 869.254016] env[62460]: DEBUG nova.compute.manager [req-09774240-dd45-4916-9b9e-e6f16f650aa5 req-4ba10221-421f-4518-b94f-befebf4a06ac service nova] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Received event network-changed-5606455d-d407-41ff-8232-841d166aa4b1 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 869.254016] env[62460]: DEBUG nova.compute.manager [req-09774240-dd45-4916-9b9e-e6f16f650aa5 req-4ba10221-421f-4518-b94f-befebf4a06ac service nova] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Refreshing instance network info cache due to event network-changed-5606455d-d407-41ff-8232-841d166aa4b1. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 869.254549] env[62460]: DEBUG oslo_concurrency.lockutils [req-09774240-dd45-4916-9b9e-e6f16f650aa5 req-4ba10221-421f-4518-b94f-befebf4a06ac service nova] Acquiring lock "refresh_cache-d7d180f7-e1a8-46c8-ba8e-ca50dac474cc" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 869.254549] env[62460]: DEBUG oslo_concurrency.lockutils [req-09774240-dd45-4916-9b9e-e6f16f650aa5 req-4ba10221-421f-4518-b94f-befebf4a06ac service nova] Acquired lock "refresh_cache-d7d180f7-e1a8-46c8-ba8e-ca50dac474cc" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.254549] env[62460]: DEBUG nova.network.neutron [req-09774240-dd45-4916-9b9e-e6f16f650aa5 req-4ba10221-421f-4518-b94f-befebf4a06ac service nova] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Refreshing network info cache for port 5606455d-d407-41ff-8232-841d166aa4b1 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 869.261166] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f299210d-270f-4dca-b764-1a4a68053d3f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.302810] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e563b137-6838-40f6-ac2f-728517c32488 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.303741] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Acquiring lock "0da70720-197c-4483-93e1-019e0f040c4c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.303963] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Lock "0da70720-197c-4483-93e1-019e0f040c4c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.311980] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Acquiring lock "fbe558f1-ce5a-4429-8833-48bdabe91071" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.312231] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Lock "fbe558f1-ce5a-4429-8833-48bdabe91071" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.314382] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aab0f6ba-23a8-44dd-841d-c94cacc22246 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.330474] env[62460]: DEBUG nova.compute.provider_tree [None req-a0a3bf88-42e5-4bc9-9abb-b76cbcd709f3 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 869.555849] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313727, 'name': CreateVM_Task} progress is 25%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.625693] env[62460]: DEBUG nova.network.neutron [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Successfully updated port: f5fba0f8-a3a5-4bcf-856c-52e2f4313154 {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 869.678184] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5770ec75-799a-4db6-97a5-b98712f6f11d tempest-ServersListShow296Test-180931441 tempest-ServersListShow296Test-180931441-project-member] Releasing lock "refresh_cache-120b7508-dc38-4659-92cb-8b3e813dae2a" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.678450] env[62460]: DEBUG nova.compute.manager [None req-5770ec75-799a-4db6-97a5-b98712f6f11d tempest-ServersListShow296Test-180931441 tempest-ServersListShow296Test-180931441-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62460) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 869.678767] env[62460]: DEBUG nova.compute.manager [None req-5770ec75-799a-4db6-97a5-b98712f6f11d tempest-ServersListShow296Test-180931441 tempest-ServersListShow296Test-180931441-project-member] [instance: 120b7508-dc38-4659-92cb-8b3e813dae2a] Skipping network deallocation for instance since networking was not requested. {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2285}} [ 869.744794] env[62460]: DEBUG oslo_vmware.api [None req-e6aa4918-c28c-4873-a5ac-d199133b8107 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313707, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.808299] env[62460]: DEBUG nova.compute.manager [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 869.824461] env[62460]: DEBUG nova.compute.manager [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 869.860734] env[62460]: ERROR nova.scheduler.client.report [None req-a0a3bf88-42e5-4bc9-9abb-b76cbcd709f3 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [req-86bef900-39d4-4834-ab87-f5783c17ce9d] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 76e8f00d-65cc-4766-bb9d-5006fb51c728. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-86bef900-39d4-4834-ab87-f5783c17ce9d"}]} [ 869.880402] env[62460]: DEBUG nova.scheduler.client.report [None req-a0a3bf88-42e5-4bc9-9abb-b76cbcd709f3 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Refreshing inventories for resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 869.884360] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-36066889-905f-461f-99f9-b5b6a17671e2 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 869.884644] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-36066889-905f-461f-99f9-b5b6a17671e2 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Deleting contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 869.884868] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-36066889-905f-461f-99f9-b5b6a17671e2 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Deleting the datastore file [datastore1] a02b4be1-91b0-4254-8d60-654885e24f6b {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 869.885203] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-19672ac4-b883-4c42-b2e4-2b5f27392510 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.892438] env[62460]: DEBUG oslo_vmware.api [None req-36066889-905f-461f-99f9-b5b6a17671e2 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Waiting for the task: (returnval){ [ 869.892438] env[62460]: value = "task-1313728" [ 869.892438] env[62460]: _type = "Task" [ 869.892438] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.896938] env[62460]: DEBUG nova.scheduler.client.report [None req-a0a3bf88-42e5-4bc9-9abb-b76cbcd709f3 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Updating ProviderTree inventory for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 869.897231] env[62460]: DEBUG nova.compute.provider_tree [None req-a0a3bf88-42e5-4bc9-9abb-b76cbcd709f3 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 869.905499] env[62460]: DEBUG oslo_vmware.api [None req-36066889-905f-461f-99f9-b5b6a17671e2 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Task: {'id': task-1313728, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.910796] env[62460]: DEBUG nova.scheduler.client.report [None req-a0a3bf88-42e5-4bc9-9abb-b76cbcd709f3 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Refreshing aggregate associations for resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728, aggregates: None {{(pid=62460) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 869.933576] env[62460]: DEBUG nova.scheduler.client.report [None req-a0a3bf88-42e5-4bc9-9abb-b76cbcd709f3 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Refreshing trait associations for resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62460) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 869.942563] env[62460]: DEBUG nova.network.neutron [req-5f0c99a6-c84f-4747-88df-c8e6086001bc req-8cba1c45-a728-4a5c-95a3-df71ac96d327 service nova] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Updated VIF entry in instance network info cache for port 530faf0d-567e-464f-99de-5f2f1d7afaf2. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 869.942656] env[62460]: DEBUG nova.network.neutron [req-5f0c99a6-c84f-4747-88df-c8e6086001bc req-8cba1c45-a728-4a5c-95a3-df71ac96d327 service nova] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Updating instance_info_cache with network_info: [{"id": "530faf0d-567e-464f-99de-5f2f1d7afaf2", "address": "fa:16:3e:7d:d3:e1", "network": {"id": "51277201-fa18-437f-95a2-e42936d56791", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1816940505-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "335cfe76a8c44686891e17906c96d158", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "84aee122-f630-43c5-9cc1-3a38d3819c82", "external-id": "nsx-vlan-transportzone-816", "segmentation_id": 816, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap530faf0d-56", "ovs_interfaceid": "530faf0d-567e-464f-99de-5f2f1d7afaf2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.945056] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e0664ea2-f133-40e3-b052-aae0e3574fee tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Acquiring lock "0f5457a0-c918-4407-9ba9-bfeeb949c6a7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.945056] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e0664ea2-f133-40e3-b052-aae0e3574fee tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Lock "0f5457a0-c918-4407-9ba9-bfeeb949c6a7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.945056] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e0664ea2-f133-40e3-b052-aae0e3574fee tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Acquiring lock "0f5457a0-c918-4407-9ba9-bfeeb949c6a7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.945056] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e0664ea2-f133-40e3-b052-aae0e3574fee tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Lock "0f5457a0-c918-4407-9ba9-bfeeb949c6a7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.945056] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e0664ea2-f133-40e3-b052-aae0e3574fee tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Lock "0f5457a0-c918-4407-9ba9-bfeeb949c6a7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.948790] env[62460]: INFO nova.compute.manager [None req-e0664ea2-f133-40e3-b052-aae0e3574fee tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] Terminating instance [ 869.950855] env[62460]: DEBUG nova.compute.manager [None req-e0664ea2-f133-40e3-b052-aae0e3574fee tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 869.951095] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-e0664ea2-f133-40e3-b052-aae0e3574fee tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 869.951996] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c81fac3-1711-4b2f-8848-3b5ca114d76f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.962355] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0664ea2-f133-40e3-b052-aae0e3574fee tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 869.963350] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2a38f81a-f1ee-4e37-a48c-5df2255b5eea {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.974647] env[62460]: DEBUG oslo_vmware.api [None req-e0664ea2-f133-40e3-b052-aae0e3574fee tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Waiting for the task: (returnval){ [ 869.974647] env[62460]: value = "task-1313729" [ 869.974647] env[62460]: _type = "Task" [ 869.974647] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.984945] env[62460]: DEBUG oslo_vmware.api [None req-e0664ea2-f133-40e3-b052-aae0e3574fee tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Task: {'id': task-1313729, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.025423] env[62460]: DEBUG oslo_concurrency.lockutils [None req-02a178f4-1158-4294-a7b0-5612d682103c tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "d7d180f7-e1a8-46c8-ba8e-ca50dac474cc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.045615] env[62460]: DEBUG nova.network.neutron [req-09774240-dd45-4916-9b9e-e6f16f650aa5 req-4ba10221-421f-4518-b94f-befebf4a06ac service nova] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Updated VIF entry in instance network info cache for port 5606455d-d407-41ff-8232-841d166aa4b1. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 870.045615] env[62460]: DEBUG nova.network.neutron [req-09774240-dd45-4916-9b9e-e6f16f650aa5 req-4ba10221-421f-4518-b94f-befebf4a06ac service nova] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Updating instance_info_cache with network_info: [{"id": "5606455d-d407-41ff-8232-841d166aa4b1", "address": "fa:16:3e:6d:35:05", "network": {"id": "de789887-9cf9-4614-a7e1-dab430aaabaf", "bridge": null, "label": "tempest-DeleteServersTestJSON-656400207-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0da056d93bdf40c39d6e82e457727ff6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap5606455d-d4", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.056048] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313727, 'name': CreateVM_Task} progress is 99%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.136783] env[62460]: DEBUG oslo_concurrency.lockutils [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "refresh_cache-fd03509c-5471-48bb-8150-d29531dca848" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.137102] env[62460]: DEBUG oslo_concurrency.lockutils [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquired lock "refresh_cache-fd03509c-5471-48bb-8150-d29531dca848" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.137285] env[62460]: DEBUG nova.network.neutron [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 870.238906] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfb28e64-5897-48a1-8b66-046fd3224f7d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.251428] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b17597f0-c3a2-4e3e-85e1-e919d329a78c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.254179] env[62460]: DEBUG oslo_vmware.api [None req-e6aa4918-c28c-4873-a5ac-d199133b8107 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313707, 'name': ReconfigVM_Task, 'duration_secs': 5.799824} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.254810] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e6aa4918-c28c-4873-a5ac-d199133b8107 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Releasing lock "7e6ff902-4a04-43d5-9014-38c4ec88efc4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.255053] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-e6aa4918-c28c-4873-a5ac-d199133b8107 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Reconfigured VM to detach interface {{(pid=62460) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 870.289938] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb10b2e6-ffbe-42cd-8c80-9b4a2fb87b46 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.298674] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e603a425-f32f-40d0-be3f-c5d6d6ffb2eb {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.311992] env[62460]: DEBUG nova.compute.provider_tree [None req-a0a3bf88-42e5-4bc9-9abb-b76cbcd709f3 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 870.334025] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.351517] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.402587] env[62460]: DEBUG oslo_vmware.api [None req-36066889-905f-461f-99f9-b5b6a17671e2 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Task: {'id': task-1313728, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.210805} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.402878] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-36066889-905f-461f-99f9-b5b6a17671e2 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 870.403090] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-36066889-905f-461f-99f9-b5b6a17671e2 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Deleted contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 870.403274] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-36066889-905f-461f-99f9-b5b6a17671e2 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 870.403452] env[62460]: INFO nova.compute.manager [None req-36066889-905f-461f-99f9-b5b6a17671e2 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Took 3.50 seconds to destroy the instance on the hypervisor. [ 870.403704] env[62460]: DEBUG oslo.service.loopingcall [None req-36066889-905f-461f-99f9-b5b6a17671e2 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 870.403935] env[62460]: DEBUG nova.compute.manager [-] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 870.403997] env[62460]: DEBUG nova.network.neutron [-] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 870.448641] env[62460]: DEBUG oslo_concurrency.lockutils [req-5f0c99a6-c84f-4747-88df-c8e6086001bc req-8cba1c45-a728-4a5c-95a3-df71ac96d327 service nova] Releasing lock "refresh_cache-3e10af70-db52-4d4a-bb92-821a05dcbab6" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.448951] env[62460]: DEBUG nova.compute.manager [req-5f0c99a6-c84f-4747-88df-c8e6086001bc req-8cba1c45-a728-4a5c-95a3-df71ac96d327 service nova] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] Received event network-vif-deleted-7b34ac0d-3fb1-4b67-8ef9-9a3a64d990db {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 870.484881] env[62460]: DEBUG oslo_vmware.api [None req-e0664ea2-f133-40e3-b052-aae0e3574fee tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Task: {'id': task-1313729, 'name': PowerOffVM_Task, 'duration_secs': 0.254644} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.485191] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0664ea2-f133-40e3-b052-aae0e3574fee tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 870.485375] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-e0664ea2-f133-40e3-b052-aae0e3574fee tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 870.485650] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-71dfd5ed-77bf-4a0f-a57f-5c38e54dde51 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.549351] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-e0664ea2-f133-40e3-b052-aae0e3574fee tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 870.549656] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-e0664ea2-f133-40e3-b052-aae0e3574fee tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] Deleting contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 870.549853] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0664ea2-f133-40e3-b052-aae0e3574fee tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Deleting the datastore file [datastore2] 0f5457a0-c918-4407-9ba9-bfeeb949c6a7 {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 870.553666] env[62460]: DEBUG oslo_concurrency.lockutils [req-09774240-dd45-4916-9b9e-e6f16f650aa5 req-4ba10221-421f-4518-b94f-befebf4a06ac service nova] Releasing lock "refresh_cache-d7d180f7-e1a8-46c8-ba8e-ca50dac474cc" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.554834] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d5635d0d-d0b8-4797-a8df-c97d4c424b3f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.556904] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313727, 'name': CreateVM_Task, 'duration_secs': 1.090271} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.557192] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 870.558094] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.558265] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.558588] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 870.559125] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6372edb1-81cb-4efd-836c-7141ee909e9b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.562467] env[62460]: DEBUG oslo_vmware.api [None req-e0664ea2-f133-40e3-b052-aae0e3574fee tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Waiting for the task: (returnval){ [ 870.562467] env[62460]: value = "task-1313731" [ 870.562467] env[62460]: _type = "Task" [ 870.562467] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.568024] env[62460]: DEBUG oslo_vmware.api [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Waiting for the task: (returnval){ [ 870.568024] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]5207a19d-124b-b036-af49-311c3a293b26" [ 870.568024] env[62460]: _type = "Task" [ 870.568024] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.575613] env[62460]: DEBUG oslo_vmware.api [None req-e0664ea2-f133-40e3-b052-aae0e3574fee tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Task: {'id': task-1313731, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.581174] env[62460]: DEBUG oslo_vmware.api [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5207a19d-124b-b036-af49-311c3a293b26, 'name': SearchDatastore_Task, 'duration_secs': 0.011125} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.581562] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.581853] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 870.582127] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.582286] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.582493] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 870.582775] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d3700965-f4f4-4e5e-9022-e7ba72e0e5f3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.591836] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 870.591972] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 870.592736] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7be878f7-1a19-4e9e-8bdd-c7c0ea436bef {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.598240] env[62460]: DEBUG oslo_vmware.api [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Waiting for the task: (returnval){ [ 870.598240] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]5215e655-50bf-9070-8118-c1c40fda1453" [ 870.598240] env[62460]: _type = "Task" [ 870.598240] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.609272] env[62460]: DEBUG oslo_vmware.api [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5215e655-50bf-9070-8118-c1c40fda1453, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.690221] env[62460]: DEBUG nova.network.neutron [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 870.708552] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5770ec75-799a-4db6-97a5-b98712f6f11d tempest-ServersListShow296Test-180931441 tempest-ServersListShow296Test-180931441-project-member] Lock "120b7508-dc38-4659-92cb-8b3e813dae2a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.258s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.853023] env[62460]: DEBUG nova.scheduler.client.report [None req-a0a3bf88-42e5-4bc9-9abb-b76cbcd709f3 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Updated inventory for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with generation 90 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 870.853023] env[62460]: DEBUG nova.compute.provider_tree [None req-a0a3bf88-42e5-4bc9-9abb-b76cbcd709f3 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Updating resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 generation from 90 to 91 during operation: update_inventory {{(pid=62460) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 870.853023] env[62460]: DEBUG nova.compute.provider_tree [None req-a0a3bf88-42e5-4bc9-9abb-b76cbcd709f3 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 871.076213] env[62460]: DEBUG oslo_vmware.api [None req-e0664ea2-f133-40e3-b052-aae0e3574fee tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Task: {'id': task-1313731, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.177048} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.076213] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0664ea2-f133-40e3-b052-aae0e3574fee tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 871.076213] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-e0664ea2-f133-40e3-b052-aae0e3574fee tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] Deleted contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 871.076213] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-e0664ea2-f133-40e3-b052-aae0e3574fee tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 871.076213] env[62460]: INFO nova.compute.manager [None req-e0664ea2-f133-40e3-b052-aae0e3574fee tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] Took 1.12 seconds to destroy the instance on the hypervisor. [ 871.076213] env[62460]: DEBUG oslo.service.loopingcall [None req-e0664ea2-f133-40e3-b052-aae0e3574fee tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 871.076213] env[62460]: DEBUG nova.compute.manager [-] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 871.076213] env[62460]: DEBUG nova.network.neutron [-] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 871.082290] env[62460]: DEBUG nova.network.neutron [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Updating instance_info_cache with network_info: [{"id": "f5fba0f8-a3a5-4bcf-856c-52e2f4313154", "address": "fa:16:3e:dd:7e:4d", "network": {"id": "3c49720c-94c9-4e2f-bbb7-4266cc97faee", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1650281790-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfa5a65b0f614e769de5b3aa77bf869e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5fba0f8-a3", "ovs_interfaceid": "f5fba0f8-a3a5-4bcf-856c-52e2f4313154", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.109090] env[62460]: DEBUG oslo_vmware.api [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5215e655-50bf-9070-8118-c1c40fda1453, 'name': SearchDatastore_Task, 'duration_secs': 0.00976} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.110072] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9a3b68ae-324e-4eb7-981a-674334ea6ad1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.116262] env[62460]: DEBUG oslo_vmware.api [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Waiting for the task: (returnval){ [ 871.116262] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52544d9b-b5cf-416f-713c-0c1204e30e20" [ 871.116262] env[62460]: _type = "Task" [ 871.116262] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.125778] env[62460]: DEBUG oslo_vmware.api [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52544d9b-b5cf-416f-713c-0c1204e30e20, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.145333] env[62460]: DEBUG nova.compute.manager [req-5c0b0fac-2eca-4576-8d67-8759099a222b req-faec555b-2ab9-45c3-a528-76085e9ffab5 service nova] [instance: fd03509c-5471-48bb-8150-d29531dca848] Received event network-vif-plugged-f5fba0f8-a3a5-4bcf-856c-52e2f4313154 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 871.145730] env[62460]: DEBUG oslo_concurrency.lockutils [req-5c0b0fac-2eca-4576-8d67-8759099a222b req-faec555b-2ab9-45c3-a528-76085e9ffab5 service nova] Acquiring lock "fd03509c-5471-48bb-8150-d29531dca848-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.146089] env[62460]: DEBUG oslo_concurrency.lockutils [req-5c0b0fac-2eca-4576-8d67-8759099a222b req-faec555b-2ab9-45c3-a528-76085e9ffab5 service nova] Lock "fd03509c-5471-48bb-8150-d29531dca848-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.146431] env[62460]: DEBUG oslo_concurrency.lockutils [req-5c0b0fac-2eca-4576-8d67-8759099a222b req-faec555b-2ab9-45c3-a528-76085e9ffab5 service nova] Lock "fd03509c-5471-48bb-8150-d29531dca848-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.146790] env[62460]: DEBUG nova.compute.manager [req-5c0b0fac-2eca-4576-8d67-8759099a222b req-faec555b-2ab9-45c3-a528-76085e9ffab5 service nova] [instance: fd03509c-5471-48bb-8150-d29531dca848] No waiting events found dispatching network-vif-plugged-f5fba0f8-a3a5-4bcf-856c-52e2f4313154 {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 871.147242] env[62460]: WARNING nova.compute.manager [req-5c0b0fac-2eca-4576-8d67-8759099a222b req-faec555b-2ab9-45c3-a528-76085e9ffab5 service nova] [instance: fd03509c-5471-48bb-8150-d29531dca848] Received unexpected event network-vif-plugged-f5fba0f8-a3a5-4bcf-856c-52e2f4313154 for instance with vm_state building and task_state spawning. [ 871.147554] env[62460]: DEBUG nova.compute.manager [req-5c0b0fac-2eca-4576-8d67-8759099a222b req-faec555b-2ab9-45c3-a528-76085e9ffab5 service nova] [instance: fd03509c-5471-48bb-8150-d29531dca848] Received event network-changed-f5fba0f8-a3a5-4bcf-856c-52e2f4313154 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 871.147952] env[62460]: DEBUG nova.compute.manager [req-5c0b0fac-2eca-4576-8d67-8759099a222b req-faec555b-2ab9-45c3-a528-76085e9ffab5 service nova] [instance: fd03509c-5471-48bb-8150-d29531dca848] Refreshing instance network info cache due to event network-changed-f5fba0f8-a3a5-4bcf-856c-52e2f4313154. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 871.148280] env[62460]: DEBUG oslo_concurrency.lockutils [req-5c0b0fac-2eca-4576-8d67-8759099a222b req-faec555b-2ab9-45c3-a528-76085e9ffab5 service nova] Acquiring lock "refresh_cache-fd03509c-5471-48bb-8150-d29531dca848" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.304304] env[62460]: DEBUG nova.compute.manager [req-6454e3f6-e604-4722-a965-92a093435e8d req-1c1f6db2-fe36-45c8-9110-59906c654f99 service nova] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Received event network-vif-deleted-c6936c73-cdb0-4475-b00b-de47fc29e7cd {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 871.304543] env[62460]: INFO nova.compute.manager [req-6454e3f6-e604-4722-a965-92a093435e8d req-1c1f6db2-fe36-45c8-9110-59906c654f99 service nova] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Neutron deleted interface c6936c73-cdb0-4475-b00b-de47fc29e7cd; detaching it from the instance and deleting it from the info cache [ 871.304723] env[62460]: DEBUG nova.network.neutron [req-6454e3f6-e604-4722-a965-92a093435e8d req-1c1f6db2-fe36-45c8-9110-59906c654f99 service nova] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.358302] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a0a3bf88-42e5-4bc9-9abb-b76cbcd709f3 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.897s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.360561] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a8b2d4df-1735-4f45-bc97-65d17e36a285 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.875s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.360804] env[62460]: DEBUG nova.objects.instance [None req-a8b2d4df-1735-4f45-bc97-65d17e36a285 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Lazy-loading 'resources' on Instance uuid 9d1a5830-f3c0-4d18-9338-16f7b6962c6a {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 871.384026] env[62460]: INFO nova.scheduler.client.report [None req-a0a3bf88-42e5-4bc9-9abb-b76cbcd709f3 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Deleted allocations for instance 16d40829-ba6e-4193-98b5-fff94c066bcb [ 871.497464] env[62460]: DEBUG nova.network.neutron [-] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.540138] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e6aa4918-c28c-4873-a5ac-d199133b8107 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "refresh_cache-7e6ff902-4a04-43d5-9014-38c4ec88efc4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.540416] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e6aa4918-c28c-4873-a5ac-d199133b8107 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquired lock "refresh_cache-7e6ff902-4a04-43d5-9014-38c4ec88efc4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.540555] env[62460]: DEBUG nova.network.neutron [None req-e6aa4918-c28c-4873-a5ac-d199133b8107 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 871.584353] env[62460]: DEBUG oslo_concurrency.lockutils [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Releasing lock "refresh_cache-fd03509c-5471-48bb-8150-d29531dca848" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.584353] env[62460]: DEBUG nova.compute.manager [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Instance network_info: |[{"id": "f5fba0f8-a3a5-4bcf-856c-52e2f4313154", "address": "fa:16:3e:dd:7e:4d", "network": {"id": "3c49720c-94c9-4e2f-bbb7-4266cc97faee", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1650281790-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfa5a65b0f614e769de5b3aa77bf869e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5fba0f8-a3", "ovs_interfaceid": "f5fba0f8-a3a5-4bcf-856c-52e2f4313154", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 871.584746] env[62460]: DEBUG oslo_concurrency.lockutils [req-5c0b0fac-2eca-4576-8d67-8759099a222b req-faec555b-2ab9-45c3-a528-76085e9ffab5 service nova] Acquired lock "refresh_cache-fd03509c-5471-48bb-8150-d29531dca848" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.584928] env[62460]: DEBUG nova.network.neutron [req-5c0b0fac-2eca-4576-8d67-8759099a222b req-faec555b-2ab9-45c3-a528-76085e9ffab5 service nova] [instance: fd03509c-5471-48bb-8150-d29531dca848] Refreshing network info cache for port f5fba0f8-a3a5-4bcf-856c-52e2f4313154 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 871.586209] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:dd:7e:4d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '04ccbc7a-cf8d-4ea2-8411-291a1e27df7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f5fba0f8-a3a5-4bcf-856c-52e2f4313154', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 871.597618] env[62460]: DEBUG oslo.service.loopingcall [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 871.597618] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fd03509c-5471-48bb-8150-d29531dca848] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 871.597618] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c940143b-d709-465e-8bd1-08411515270a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.618724] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 871.618724] env[62460]: value = "task-1313732" [ 871.618724] env[62460]: _type = "Task" [ 871.618724] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.630770] env[62460]: DEBUG oslo_vmware.api [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52544d9b-b5cf-416f-713c-0c1204e30e20, 'name': SearchDatastore_Task, 'duration_secs': 0.026795} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.633811] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.634073] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 3e10af70-db52-4d4a-bb92-821a05dcbab6/3e10af70-db52-4d4a-bb92-821a05dcbab6.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 871.634608] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313732, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.634823] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e4bec7e1-243d-4d3b-86fa-3014e3e82b99 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.641317] env[62460]: DEBUG oslo_vmware.api [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Waiting for the task: (returnval){ [ 871.641317] env[62460]: value = "task-1313733" [ 871.641317] env[62460]: _type = "Task" [ 871.641317] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.649555] env[62460]: DEBUG oslo_vmware.api [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Task: {'id': task-1313733, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.812485] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-72101797-cbb6-416a-a096-f0238fb7abd4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.822263] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-035a99fb-5f11-4895-a235-e76af3016266 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.849054] env[62460]: DEBUG nova.compute.manager [req-6454e3f6-e604-4722-a965-92a093435e8d req-1c1f6db2-fe36-45c8-9110-59906c654f99 service nova] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Detach interface failed, port_id=c6936c73-cdb0-4475-b00b-de47fc29e7cd, reason: Instance a02b4be1-91b0-4254-8d60-654885e24f6b could not be found. {{(pid=62460) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 871.892147] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a0a3bf88-42e5-4bc9-9abb-b76cbcd709f3 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "16d40829-ba6e-4193-98b5-fff94c066bcb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.411s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.937024] env[62460]: DEBUG nova.network.neutron [-] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.938770] env[62460]: DEBUG oslo_concurrency.lockutils [None req-f9e6de72-4edb-4918-b758-21cb807f86bc tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "7e6ff902-4a04-43d5-9014-38c4ec88efc4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.939152] env[62460]: DEBUG oslo_concurrency.lockutils [None req-f9e6de72-4edb-4918-b758-21cb807f86bc tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "7e6ff902-4a04-43d5-9014-38c4ec88efc4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.939478] env[62460]: DEBUG oslo_concurrency.lockutils [None req-f9e6de72-4edb-4918-b758-21cb807f86bc tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "7e6ff902-4a04-43d5-9014-38c4ec88efc4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.939993] env[62460]: DEBUG oslo_concurrency.lockutils [None req-f9e6de72-4edb-4918-b758-21cb807f86bc tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "7e6ff902-4a04-43d5-9014-38c4ec88efc4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.940476] env[62460]: DEBUG oslo_concurrency.lockutils [None req-f9e6de72-4edb-4918-b758-21cb807f86bc tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "7e6ff902-4a04-43d5-9014-38c4ec88efc4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.947814] env[62460]: INFO nova.compute.manager [None req-f9e6de72-4edb-4918-b758-21cb807f86bc tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Terminating instance [ 871.950562] env[62460]: DEBUG nova.compute.manager [None req-f9e6de72-4edb-4918-b758-21cb807f86bc tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 871.951033] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-f9e6de72-4edb-4918-b758-21cb807f86bc tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 871.951661] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e007ac4d-8738-401b-bfe1-3b4311d50206 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.963597] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9e6de72-4edb-4918-b758-21cb807f86bc tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 871.964011] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8010b9f5-d29e-4ff2-aa93-51ad48078a18 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.970322] env[62460]: DEBUG oslo_vmware.api [None req-f9e6de72-4edb-4918-b758-21cb807f86bc tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Waiting for the task: (returnval){ [ 871.970322] env[62460]: value = "task-1313734" [ 871.970322] env[62460]: _type = "Task" [ 871.970322] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.982642] env[62460]: DEBUG oslo_vmware.api [None req-f9e6de72-4edb-4918-b758-21cb807f86bc tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313734, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.999742] env[62460]: INFO nova.compute.manager [-] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Took 1.60 seconds to deallocate network for instance. [ 872.092050] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfbadc66-b996-4e1d-b37f-60ae34be782d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.107124] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ca1a2dd-6e42-437a-ac72-23b7a19bf404 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.142093] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3edb5d6-9c5a-44cd-ab75-ec1f9a766778 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.151925] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313732, 'name': CreateVM_Task, 'duration_secs': 0.41057} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.156085] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fd03509c-5471-48bb-8150-d29531dca848] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 872.160399] env[62460]: DEBUG oslo_concurrency.lockutils [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.160605] env[62460]: DEBUG oslo_concurrency.lockutils [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.160945] env[62460]: DEBUG oslo_concurrency.lockutils [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 872.162374] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ff036d1-a0ab-472b-9354-52b652d9c4b1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.171446] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-77a86d1f-6853-4a03-8569-a5375033fb18 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.173605] env[62460]: DEBUG oslo_vmware.api [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Task: {'id': task-1313733, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.184067] env[62460]: DEBUG nova.compute.provider_tree [None req-a8b2d4df-1735-4f45-bc97-65d17e36a285 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 872.187053] env[62460]: DEBUG oslo_vmware.api [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 872.187053] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52d3866d-af4e-d78a-f604-7c0a6b4e614a" [ 872.187053] env[62460]: _type = "Task" [ 872.187053] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.196514] env[62460]: DEBUG oslo_vmware.api [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52d3866d-af4e-d78a-f604-7c0a6b4e614a, 'name': SearchDatastore_Task, 'duration_secs': 0.016698} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.197821] env[62460]: DEBUG oslo_concurrency.lockutils [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.197821] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 872.197931] env[62460]: DEBUG oslo_concurrency.lockutils [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.198099] env[62460]: DEBUG oslo_concurrency.lockutils [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.198299] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 872.198926] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-439183a8-8d9e-4115-8092-cfc5c6cd2cd1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.209687] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 872.210481] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 872.211274] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6fa6994a-3978-4367-b020-2e04663408f0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.217451] env[62460]: DEBUG oslo_vmware.api [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 872.217451] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]521c28b1-aa14-2929-db08-137f08b52664" [ 872.217451] env[62460]: _type = "Task" [ 872.217451] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.225521] env[62460]: DEBUG oslo_vmware.api [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]521c28b1-aa14-2929-db08-137f08b52664, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.397728] env[62460]: DEBUG oslo_concurrency.lockutils [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Acquiring lock "8bcaa672-97c8-466a-812c-5723e08d0c52" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.397997] env[62460]: DEBUG oslo_concurrency.lockutils [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Lock "8bcaa672-97c8-466a-812c-5723e08d0c52" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.426115] env[62460]: DEBUG nova.network.neutron [req-5c0b0fac-2eca-4576-8d67-8759099a222b req-faec555b-2ab9-45c3-a528-76085e9ffab5 service nova] [instance: fd03509c-5471-48bb-8150-d29531dca848] Updated VIF entry in instance network info cache for port f5fba0f8-a3a5-4bcf-856c-52e2f4313154. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 872.426518] env[62460]: DEBUG nova.network.neutron [req-5c0b0fac-2eca-4576-8d67-8759099a222b req-faec555b-2ab9-45c3-a528-76085e9ffab5 service nova] [instance: fd03509c-5471-48bb-8150-d29531dca848] Updating instance_info_cache with network_info: [{"id": "f5fba0f8-a3a5-4bcf-856c-52e2f4313154", "address": "fa:16:3e:dd:7e:4d", "network": {"id": "3c49720c-94c9-4e2f-bbb7-4266cc97faee", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1650281790-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfa5a65b0f614e769de5b3aa77bf869e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5fba0f8-a3", "ovs_interfaceid": "f5fba0f8-a3a5-4bcf-856c-52e2f4313154", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.429681] env[62460]: INFO nova.network.neutron [None req-e6aa4918-c28c-4873-a5ac-d199133b8107 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Port bf0f715f-7988-45f9-81b6-c92688ff60c7 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 872.429864] env[62460]: INFO nova.network.neutron [None req-e6aa4918-c28c-4873-a5ac-d199133b8107 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Port d1397bb8-e534-405e-9921-e3e7e6914f81 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 872.430189] env[62460]: DEBUG nova.network.neutron [None req-e6aa4918-c28c-4873-a5ac-d199133b8107 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Updating instance_info_cache with network_info: [{"id": "ba38cbb0-62cf-47cb-a238-87ebc03de250", "address": "fa:16:3e:b4:93:18", "network": {"id": "d3136e32-ad55-4b73-835c-8fa5f0480767", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1511414968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53fa06008e9f43488362895e7a143700", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "054fcd1e-638e-425a-a1de-78cb188ae026", "external-id": "nsx-vlan-transportzone-658", "segmentation_id": 658, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba38cbb0-62", "ovs_interfaceid": "ba38cbb0-62cf-47cb-a238-87ebc03de250", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.444081] env[62460]: INFO nova.compute.manager [-] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] Took 1.37 seconds to deallocate network for instance. [ 872.482272] env[62460]: DEBUG oslo_vmware.api [None req-f9e6de72-4edb-4918-b758-21cb807f86bc tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313734, 'name': PowerOffVM_Task, 'duration_secs': 0.370246} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.482563] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9e6de72-4edb-4918-b758-21cb807f86bc tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 872.482742] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-f9e6de72-4edb-4918-b758-21cb807f86bc tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 872.483036] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ce8e2037-911b-480a-81bd-e9fddfdb4587 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.508539] env[62460]: DEBUG oslo_concurrency.lockutils [None req-36066889-905f-461f-99f9-b5b6a17671e2 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.591552] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-f9e6de72-4edb-4918-b758-21cb807f86bc tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 872.591925] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-f9e6de72-4edb-4918-b758-21cb807f86bc tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Deleting contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 872.592798] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-f9e6de72-4edb-4918-b758-21cb807f86bc tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Deleting the datastore file [datastore1] 7e6ff902-4a04-43d5-9014-38c4ec88efc4 {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 872.592798] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c7266b06-fe9a-413a-b6b0-353a16560201 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.613999] env[62460]: DEBUG oslo_vmware.api [None req-f9e6de72-4edb-4918-b758-21cb807f86bc tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Waiting for the task: (returnval){ [ 872.613999] env[62460]: value = "task-1313736" [ 872.613999] env[62460]: _type = "Task" [ 872.613999] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.634730] env[62460]: DEBUG oslo_vmware.api [None req-f9e6de72-4edb-4918-b758-21cb807f86bc tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313736, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.661191] env[62460]: DEBUG oslo_vmware.api [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Task: {'id': task-1313733, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.565531} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.661467] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 3e10af70-db52-4d4a-bb92-821a05dcbab6/3e10af70-db52-4d4a-bb92-821a05dcbab6.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 872.661689] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 872.661955] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-eee72b51-6529-49d6-95ff-339ede837ddd {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.670836] env[62460]: DEBUG oslo_vmware.api [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Waiting for the task: (returnval){ [ 872.670836] env[62460]: value = "task-1313737" [ 872.670836] env[62460]: _type = "Task" [ 872.670836] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.679485] env[62460]: DEBUG oslo_vmware.api [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Task: {'id': task-1313737, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.690978] env[62460]: DEBUG nova.scheduler.client.report [None req-a8b2d4df-1735-4f45-bc97-65d17e36a285 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 872.725634] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "e90e8c3a-d244-4d04-997e-c15ab1cb2009" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.725851] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "e90e8c3a-d244-4d04-997e-c15ab1cb2009" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.730951] env[62460]: DEBUG oslo_vmware.api [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]521c28b1-aa14-2929-db08-137f08b52664, 'name': SearchDatastore_Task, 'duration_secs': 0.010887} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.731887] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4eb84f78-c8b7-446e-9211-7cc787a088ee {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.736964] env[62460]: DEBUG oslo_vmware.api [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 872.736964] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]524d9370-d126-2dab-66fa-cf35c960b799" [ 872.736964] env[62460]: _type = "Task" [ 872.736964] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.745148] env[62460]: DEBUG oslo_vmware.api [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]524d9370-d126-2dab-66fa-cf35c960b799, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.901918] env[62460]: DEBUG nova.compute.manager [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 872.932782] env[62460]: DEBUG oslo_concurrency.lockutils [req-5c0b0fac-2eca-4576-8d67-8759099a222b req-faec555b-2ab9-45c3-a528-76085e9ffab5 service nova] Releasing lock "refresh_cache-fd03509c-5471-48bb-8150-d29531dca848" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.933096] env[62460]: DEBUG nova.compute.manager [req-5c0b0fac-2eca-4576-8d67-8759099a222b req-faec555b-2ab9-45c3-a528-76085e9ffab5 service nova] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Received event network-vif-deleted-bf0f715f-7988-45f9-81b6-c92688ff60c7 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 872.933297] env[62460]: INFO nova.compute.manager [req-5c0b0fac-2eca-4576-8d67-8759099a222b req-faec555b-2ab9-45c3-a528-76085e9ffab5 service nova] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Neutron deleted interface bf0f715f-7988-45f9-81b6-c92688ff60c7; detaching it from the instance and deleting it from the info cache [ 872.933633] env[62460]: DEBUG nova.network.neutron [req-5c0b0fac-2eca-4576-8d67-8759099a222b req-faec555b-2ab9-45c3-a528-76085e9ffab5 service nova] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Updating instance_info_cache with network_info: [{"id": "ba38cbb0-62cf-47cb-a238-87ebc03de250", "address": "fa:16:3e:b4:93:18", "network": {"id": "d3136e32-ad55-4b73-835c-8fa5f0480767", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1511414968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53fa06008e9f43488362895e7a143700", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "054fcd1e-638e-425a-a1de-78cb188ae026", "external-id": "nsx-vlan-transportzone-658", "segmentation_id": 658, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba38cbb0-62", "ovs_interfaceid": "ba38cbb0-62cf-47cb-a238-87ebc03de250", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "d1397bb8-e534-405e-9921-e3e7e6914f81", "address": "fa:16:3e:2e:b4:bd", "network": {"id": "d3136e32-ad55-4b73-835c-8fa5f0480767", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1511414968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53fa06008e9f43488362895e7a143700", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "054fcd1e-638e-425a-a1de-78cb188ae026", "external-id": "nsx-vlan-transportzone-658", "segmentation_id": 658, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd1397bb8-e5", "ovs_interfaceid": "d1397bb8-e534-405e-9921-e3e7e6914f81", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.935122] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e6aa4918-c28c-4873-a5ac-d199133b8107 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Releasing lock "refresh_cache-7e6ff902-4a04-43d5-9014-38c4ec88efc4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.949993] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e0664ea2-f133-40e3-b052-aae0e3574fee tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.124043] env[62460]: DEBUG oslo_vmware.api [None req-f9e6de72-4edb-4918-b758-21cb807f86bc tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313736, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.177963} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.124178] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-f9e6de72-4edb-4918-b758-21cb807f86bc tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 873.124352] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-f9e6de72-4edb-4918-b758-21cb807f86bc tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Deleted contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 873.124559] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-f9e6de72-4edb-4918-b758-21cb807f86bc tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 873.124746] env[62460]: INFO nova.compute.manager [None req-f9e6de72-4edb-4918-b758-21cb807f86bc tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Took 1.17 seconds to destroy the instance on the hypervisor. [ 873.124995] env[62460]: DEBUG oslo.service.loopingcall [None req-f9e6de72-4edb-4918-b758-21cb807f86bc tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 873.125214] env[62460]: DEBUG nova.compute.manager [-] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 873.125309] env[62460]: DEBUG nova.network.neutron [-] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 873.171264] env[62460]: DEBUG nova.compute.manager [req-78450dd7-aa80-4c22-8bda-f673c46009f1 req-b98126d4-d1f6-45fe-9f61-45c52a6f9344 service nova] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Received event network-vif-deleted-d1397bb8-e534-405e-9921-e3e7e6914f81 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 873.171546] env[62460]: INFO nova.compute.manager [req-78450dd7-aa80-4c22-8bda-f673c46009f1 req-b98126d4-d1f6-45fe-9f61-45c52a6f9344 service nova] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Neutron deleted interface d1397bb8-e534-405e-9921-e3e7e6914f81; detaching it from the instance and deleting it from the info cache [ 873.171889] env[62460]: DEBUG nova.network.neutron [req-78450dd7-aa80-4c22-8bda-f673c46009f1 req-b98126d4-d1f6-45fe-9f61-45c52a6f9344 service nova] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Updating instance_info_cache with network_info: [{"id": "ba38cbb0-62cf-47cb-a238-87ebc03de250", "address": "fa:16:3e:b4:93:18", "network": {"id": "d3136e32-ad55-4b73-835c-8fa5f0480767", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1511414968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53fa06008e9f43488362895e7a143700", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "054fcd1e-638e-425a-a1de-78cb188ae026", "external-id": "nsx-vlan-transportzone-658", "segmentation_id": 658, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba38cbb0-62", "ovs_interfaceid": "ba38cbb0-62cf-47cb-a238-87ebc03de250", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.174145] env[62460]: DEBUG neutronclient.v2_0.client [-] Error message: {"NeutronError": {"type": "PortNotFound", "message": "Port d1397bb8-e534-405e-9921-e3e7e6914f81 could not be found.", "detail": ""}} {{(pid=62460) _handle_fault_response /opt/stack/data/venv/lib/python3.10/site-packages/neutronclient/v2_0/client.py:262}} [ 873.175274] env[62460]: DEBUG nova.network.neutron [-] Unable to show port d1397bb8-e534-405e-9921-e3e7e6914f81 as it no longer exists. {{(pid=62460) _unbind_ports /opt/stack/nova/nova/network/neutron.py:666}} [ 873.188806] env[62460]: DEBUG oslo_vmware.api [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Task: {'id': task-1313737, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062967} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.189943] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 873.191228] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d45d31a-7270-4e1d-bb90-c192fc2bf058 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.196830] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a8b2d4df-1735-4f45-bc97-65d17e36a285 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.836s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.200226] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4db91be9-d7a0-4b00-8292-14baeb7ea3f2 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.477s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.200615] env[62460]: DEBUG nova.objects.instance [None req-4db91be9-d7a0-4b00-8292-14baeb7ea3f2 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Lazy-loading 'resources' on Instance uuid 49477d35-92ea-4f9d-8333-fc23144b7dfc {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 873.236052] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Reconfiguring VM instance instance-0000004d to attach disk [datastore1] 3e10af70-db52-4d4a-bb92-821a05dcbab6/3e10af70-db52-4d4a-bb92-821a05dcbab6.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 873.237622] env[62460]: INFO nova.scheduler.client.report [None req-a8b2d4df-1735-4f45-bc97-65d17e36a285 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Deleted allocations for instance 9d1a5830-f3c0-4d18-9338-16f7b6962c6a [ 873.242201] env[62460]: DEBUG nova.compute.manager [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 873.245272] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-14747b75-ff91-4b23-9c57-7f328fa0e7f0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.290855] env[62460]: DEBUG oslo_vmware.api [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]524d9370-d126-2dab-66fa-cf35c960b799, 'name': SearchDatastore_Task, 'duration_secs': 0.009197} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.292680] env[62460]: DEBUG oslo_concurrency.lockutils [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.293103] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] fd03509c-5471-48bb-8150-d29531dca848/fd03509c-5471-48bb-8150-d29531dca848.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 873.294565] env[62460]: DEBUG oslo_vmware.api [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Waiting for the task: (returnval){ [ 873.294565] env[62460]: value = "task-1313738" [ 873.294565] env[62460]: _type = "Task" [ 873.294565] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.294565] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6617f497-6a6b-4568-9ff3-d39edeaaf21a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.308615] env[62460]: DEBUG oslo_vmware.api [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Task: {'id': task-1313738, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.310357] env[62460]: DEBUG oslo_vmware.api [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 873.310357] env[62460]: value = "task-1313739" [ 873.310357] env[62460]: _type = "Task" [ 873.310357] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.321535] env[62460]: DEBUG oslo_vmware.api [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313739, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.340456] env[62460]: DEBUG nova.compute.manager [req-fb197868-3e5e-4c33-a601-324507e89f67 req-dee8bf1f-ad83-420b-864e-bf36463eed2d service nova] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] Received event network-vif-deleted-948f172e-f4a4-484c-b845-ccb61e37c1b2 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 873.421149] env[62460]: DEBUG oslo_concurrency.lockutils [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.438054] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-218f48e5-ef05-4f1d-a9e6-8ce41be43f1c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.443998] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e6aa4918-c28c-4873-a5ac-d199133b8107 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "interface-7e6ff902-4a04-43d5-9014-38c4ec88efc4-bf0f715f-7988-45f9-81b6-c92688ff60c7" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.807s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.448573] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99b24f01-a571-4254-82eb-a833eff6b88c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.477300] env[62460]: DEBUG nova.compute.manager [req-5c0b0fac-2eca-4576-8d67-8759099a222b req-faec555b-2ab9-45c3-a528-76085e9ffab5 service nova] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Detach interface failed, port_id=bf0f715f-7988-45f9-81b6-c92688ff60c7, reason: Instance 7e6ff902-4a04-43d5-9014-38c4ec88efc4 could not be found. {{(pid=62460) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 873.675067] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ef2c6fe4-f285-4942-be2d-31d9ac847aa9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.693038] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c93dcc7-d6b8-4cd1-bfa8-a1daa5b558e2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.728803] env[62460]: DEBUG nova.compute.manager [req-78450dd7-aa80-4c22-8bda-f673c46009f1 req-b98126d4-d1f6-45fe-9f61-45c52a6f9344 service nova] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Detach interface failed, port_id=d1397bb8-e534-405e-9921-e3e7e6914f81, reason: Instance 7e6ff902-4a04-43d5-9014-38c4ec88efc4 could not be found. {{(pid=62460) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 873.786092] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a8b2d4df-1735-4f45-bc97-65d17e36a285 tempest-ServerShowV254Test-52212450 tempest-ServerShowV254Test-52212450-project-member] Lock "9d1a5830-f3c0-4d18-9338-16f7b6962c6a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.077s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.791125] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.816738] env[62460]: DEBUG oslo_vmware.api [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Task: {'id': task-1313738, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.826271] env[62460]: DEBUG oslo_vmware.api [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313739, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.996520] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b419a222-27d4-4e3d-ba37-0b3de05d8fd7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.004681] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d8edfaf-c06b-43f4-ac3d-cabdb3a4dc37 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.035438] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1e366ad-6063-471a-916d-347aa6af7634 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.044251] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13b053ed-8b01-428e-860d-95df92c68056 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.058194] env[62460]: DEBUG nova.compute.provider_tree [None req-4db91be9-d7a0-4b00-8292-14baeb7ea3f2 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 874.278619] env[62460]: DEBUG nova.network.neutron [-] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.308760] env[62460]: DEBUG oslo_vmware.api [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Task: {'id': task-1313738, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.323732] env[62460]: DEBUG oslo_vmware.api [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313739, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.612025} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.324054] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] fd03509c-5471-48bb-8150-d29531dca848/fd03509c-5471-48bb-8150-d29531dca848.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 874.324251] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 874.324561] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1c1d34c8-f499-4b55-8e20-16baa4a6a38e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.331538] env[62460]: DEBUG oslo_vmware.api [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 874.331538] env[62460]: value = "task-1313740" [ 874.331538] env[62460]: _type = "Task" [ 874.331538] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.340794] env[62460]: DEBUG oslo_vmware.api [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313740, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.583165] env[62460]: ERROR nova.scheduler.client.report [None req-4db91be9-d7a0-4b00-8292-14baeb7ea3f2 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] [req-803f36fa-11e3-4f54-8092-2e36182dbb9d] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 76e8f00d-65cc-4766-bb9d-5006fb51c728. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-803f36fa-11e3-4f54-8092-2e36182dbb9d"}]} [ 874.598659] env[62460]: DEBUG nova.scheduler.client.report [None req-4db91be9-d7a0-4b00-8292-14baeb7ea3f2 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Refreshing inventories for resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 874.614569] env[62460]: DEBUG nova.scheduler.client.report [None req-4db91be9-d7a0-4b00-8292-14baeb7ea3f2 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Updating ProviderTree inventory for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 874.614808] env[62460]: DEBUG nova.compute.provider_tree [None req-4db91be9-d7a0-4b00-8292-14baeb7ea3f2 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 874.629738] env[62460]: DEBUG nova.scheduler.client.report [None req-4db91be9-d7a0-4b00-8292-14baeb7ea3f2 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Refreshing aggregate associations for resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728, aggregates: None {{(pid=62460) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 874.652641] env[62460]: DEBUG nova.scheduler.client.report [None req-4db91be9-d7a0-4b00-8292-14baeb7ea3f2 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Refreshing trait associations for resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62460) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 874.781393] env[62460]: INFO nova.compute.manager [-] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Took 1.66 seconds to deallocate network for instance. [ 874.809619] env[62460]: DEBUG oslo_vmware.api [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Task: {'id': task-1313738, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.825472] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-045618ff-59e5-412e-ad27-792eaf3cdd0d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.835274] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20ecc272-95ec-46af-b92d-db03425b893c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.842983] env[62460]: DEBUG oslo_vmware.api [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313740, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060007} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.865965] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 874.867180] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73b8b704-4da4-46f6-97f5-c59d3606b65c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.869959] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c242a111-31b8-4f6b-a6f3-0818b466e0be {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.890706] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Reconfiguring VM instance instance-0000004e to attach disk [datastore2] fd03509c-5471-48bb-8150-d29531dca848/fd03509c-5471-48bb-8150-d29531dca848.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 874.892647] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0348f458-a6c1-4350-af03-f6f5106247dd {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.907584] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e0a54c7-198e-407f-b439-e654c06917fc {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.920959] env[62460]: DEBUG nova.compute.provider_tree [None req-4db91be9-d7a0-4b00-8292-14baeb7ea3f2 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 874.923265] env[62460]: DEBUG oslo_vmware.api [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 874.923265] env[62460]: value = "task-1313741" [ 874.923265] env[62460]: _type = "Task" [ 874.923265] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.931008] env[62460]: DEBUG oslo_vmware.api [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313741, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.030236] env[62460]: DEBUG oslo_concurrency.lockutils [None req-15b5a62f-433a-41dd-8a74-63304b656114 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Acquiring lock "5b6e8205-003e-49c0-a73d-be2e032a8272" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.030504] env[62460]: DEBUG oslo_concurrency.lockutils [None req-15b5a62f-433a-41dd-8a74-63304b656114 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lock "5b6e8205-003e-49c0-a73d-be2e032a8272" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.290780] env[62460]: DEBUG oslo_concurrency.lockutils [None req-f9e6de72-4edb-4918-b758-21cb807f86bc tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.308196] env[62460]: DEBUG oslo_vmware.api [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Task: {'id': task-1313738, 'name': ReconfigVM_Task, 'duration_secs': 2.018833} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.308512] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Reconfigured VM instance instance-0000004d to attach disk [datastore1] 3e10af70-db52-4d4a-bb92-821a05dcbab6/3e10af70-db52-4d4a-bb92-821a05dcbab6.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 875.309174] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4e6c4534-7386-42d1-85d7-15419bca8c84 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.315079] env[62460]: DEBUG oslo_vmware.api [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Waiting for the task: (returnval){ [ 875.315079] env[62460]: value = "task-1313742" [ 875.315079] env[62460]: _type = "Task" [ 875.315079] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.322989] env[62460]: DEBUG oslo_vmware.api [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Task: {'id': task-1313742, 'name': Rename_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.366008] env[62460]: DEBUG nova.compute.manager [req-4c1535b3-665d-4ba7-ae82-94ba9b53c537 req-94adc27a-6297-470b-8572-3887d8b0ce33 service nova] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Received event network-vif-deleted-ba38cbb0-62cf-47cb-a238-87ebc03de250 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 875.435934] env[62460]: DEBUG oslo_vmware.api [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313741, 'name': ReconfigVM_Task, 'duration_secs': 0.297099} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.436246] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Reconfigured VM instance instance-0000004e to attach disk [datastore2] fd03509c-5471-48bb-8150-d29531dca848/fd03509c-5471-48bb-8150-d29531dca848.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 875.436985] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-70682139-4413-4394-8ac5-8a5ad87d69a5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.442757] env[62460]: DEBUG oslo_vmware.api [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 875.442757] env[62460]: value = "task-1313743" [ 875.442757] env[62460]: _type = "Task" [ 875.442757] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.451559] env[62460]: DEBUG oslo_vmware.api [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313743, 'name': Rename_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.453306] env[62460]: DEBUG nova.scheduler.client.report [None req-4db91be9-d7a0-4b00-8292-14baeb7ea3f2 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Updated inventory for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with generation 93 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 875.453546] env[62460]: DEBUG nova.compute.provider_tree [None req-4db91be9-d7a0-4b00-8292-14baeb7ea3f2 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Updating resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 generation from 93 to 94 during operation: update_inventory {{(pid=62460) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 875.453731] env[62460]: DEBUG nova.compute.provider_tree [None req-4db91be9-d7a0-4b00-8292-14baeb7ea3f2 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 875.533442] env[62460]: DEBUG nova.compute.utils [None req-15b5a62f-433a-41dd-8a74-63304b656114 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 875.824917] env[62460]: DEBUG oslo_vmware.api [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Task: {'id': task-1313742, 'name': Rename_Task, 'duration_secs': 0.136986} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.825225] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 875.825478] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f88da2dd-f735-4d8f-b718-371edc24222a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.832158] env[62460]: DEBUG oslo_vmware.api [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Waiting for the task: (returnval){ [ 875.832158] env[62460]: value = "task-1313744" [ 875.832158] env[62460]: _type = "Task" [ 875.832158] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.839500] env[62460]: DEBUG oslo_vmware.api [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Task: {'id': task-1313744, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.954628] env[62460]: DEBUG oslo_vmware.api [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313743, 'name': Rename_Task, 'duration_secs': 0.134775} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.954628] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 875.954776] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-14b8ffeb-042c-445e-ba28-54a5b68e87c5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.958456] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4db91be9-d7a0-4b00-8292-14baeb7ea3f2 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.758s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.961908] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b696d9d5-b153-4d81-a79f-4ae255c22926 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.064s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.961908] env[62460]: DEBUG nova.objects.instance [None req-b696d9d5-b153-4d81-a79f-4ae255c22926 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Lazy-loading 'resources' on Instance uuid 3ada3516-3147-4566-a46a-1cb29cf880d0 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 875.963137] env[62460]: DEBUG oslo_vmware.api [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 875.963137] env[62460]: value = "task-1313745" [ 875.963137] env[62460]: _type = "Task" [ 875.963137] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.971216] env[62460]: DEBUG oslo_vmware.api [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313745, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.986796] env[62460]: INFO nova.scheduler.client.report [None req-4db91be9-d7a0-4b00-8292-14baeb7ea3f2 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Deleted allocations for instance 49477d35-92ea-4f9d-8333-fc23144b7dfc [ 876.036083] env[62460]: DEBUG oslo_concurrency.lockutils [None req-15b5a62f-433a-41dd-8a74-63304b656114 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lock "5b6e8205-003e-49c0-a73d-be2e032a8272" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.005s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.343683] env[62460]: DEBUG oslo_vmware.api [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Task: {'id': task-1313744, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.475294] env[62460]: DEBUG oslo_vmware.api [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313745, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.495611] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4db91be9-d7a0-4b00-8292-14baeb7ea3f2 tempest-ServerTagsTestJSON-170041924 tempest-ServerTagsTestJSON-170041924-project-member] Lock "49477d35-92ea-4f9d-8333-fc23144b7dfc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.772s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.662341] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c4941b7-cf19-465a-a16e-b4de25996164 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.671204] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ffcd2b2-e059-4042-80ee-4ae36d85040f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.703812] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0227360a-b264-44bb-af23-0ccc808070a0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.712124] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbb54b4f-67be-4416-b78d-faa677c281db {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.725843] env[62460]: DEBUG nova.compute.provider_tree [None req-b696d9d5-b153-4d81-a79f-4ae255c22926 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 876.845595] env[62460]: DEBUG oslo_vmware.api [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Task: {'id': task-1313744, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.979056] env[62460]: DEBUG oslo_vmware.api [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313745, 'name': PowerOnVM_Task, 'duration_secs': 0.992815} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.979481] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 876.979776] env[62460]: INFO nova.compute.manager [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Took 8.02 seconds to spawn the instance on the hypervisor. [ 876.980013] env[62460]: DEBUG nova.compute.manager [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 876.980856] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66966416-02fe-4177-8d6e-4b8690feec2c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.109691] env[62460]: DEBUG oslo_concurrency.lockutils [None req-15b5a62f-433a-41dd-8a74-63304b656114 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Acquiring lock "5b6e8205-003e-49c0-a73d-be2e032a8272" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.110102] env[62460]: DEBUG oslo_concurrency.lockutils [None req-15b5a62f-433a-41dd-8a74-63304b656114 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lock "5b6e8205-003e-49c0-a73d-be2e032a8272" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.110435] env[62460]: INFO nova.compute.manager [None req-15b5a62f-433a-41dd-8a74-63304b656114 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Attaching volume a05c7ae1-7cd9-447e-8453-15a2bcde137e to /dev/sdb [ 877.141608] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cfd73f6-6d35-4bfb-97e1-3f3440199ea2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.148793] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a9dc278-cc3d-4a02-b9ea-59576f8308d0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.162814] env[62460]: DEBUG nova.virt.block_device [None req-15b5a62f-433a-41dd-8a74-63304b656114 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Updating existing volume attachment record: 2c120ff1-d732-4cea-ad41-30e304d75ec3 {{(pid=62460) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 877.229530] env[62460]: DEBUG nova.scheduler.client.report [None req-b696d9d5-b153-4d81-a79f-4ae255c22926 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 877.348898] env[62460]: DEBUG oslo_vmware.api [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Task: {'id': task-1313744, 'name': PowerOnVM_Task, 'duration_secs': 1.035895} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.349345] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 877.349705] env[62460]: INFO nova.compute.manager [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Took 11.82 seconds to spawn the instance on the hypervisor. [ 877.351471] env[62460]: DEBUG nova.compute.manager [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 877.352441] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9731a91-29c7-4389-892c-f29ca1122866 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.504032] env[62460]: INFO nova.compute.manager [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Took 29.73 seconds to build instance. [ 877.736104] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b696d9d5-b153-4d81-a79f-4ae255c22926 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.773s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.736965] env[62460]: DEBUG oslo_concurrency.lockutils [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.572s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.738057] env[62460]: DEBUG nova.objects.instance [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lazy-loading 'resources' on Instance uuid d7d180f7-e1a8-46c8-ba8e-ca50dac474cc {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 877.762208] env[62460]: INFO nova.scheduler.client.report [None req-b696d9d5-b153-4d81-a79f-4ae255c22926 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Deleted allocations for instance 3ada3516-3147-4566-a46a-1cb29cf880d0 [ 877.874074] env[62460]: INFO nova.compute.manager [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Took 38.03 seconds to build instance. [ 878.007083] env[62460]: DEBUG oslo_concurrency.lockutils [None req-44231813-dadd-481c-9314-c2a0cb07537c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "fd03509c-5471-48bb-8150-d29531dca848" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.234s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.240576] env[62460]: DEBUG nova.objects.instance [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lazy-loading 'numa_topology' on Instance uuid d7d180f7-e1a8-46c8-ba8e-ca50dac474cc {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 878.272947] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b696d9d5-b153-4d81-a79f-4ae255c22926 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Lock "3ada3516-3147-4566-a46a-1cb29cf880d0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.354s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.376246] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1799ad17-e0b2-46be-bc9f-37d274a6530d tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Lock "3e10af70-db52-4d4a-bb92-821a05dcbab6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 56.642s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.510636] env[62460]: INFO nova.compute.manager [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Rebuilding instance [ 878.561701] env[62460]: DEBUG nova.compute.manager [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 878.562604] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8403de28-5641-4813-b45e-b96a80d6d703 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.743629] env[62460]: DEBUG nova.objects.base [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=62460) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 878.873924] env[62460]: DEBUG nova.compute.manager [req-ae7ec45c-8c02-43bd-b5de-48648b1b557f req-a43aefc6-d189-4dfd-953f-8302a2ac26ea service nova] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Received event network-changed-530faf0d-567e-464f-99de-5f2f1d7afaf2 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 878.874164] env[62460]: DEBUG nova.compute.manager [req-ae7ec45c-8c02-43bd-b5de-48648b1b557f req-a43aefc6-d189-4dfd-953f-8302a2ac26ea service nova] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Refreshing instance network info cache due to event network-changed-530faf0d-567e-464f-99de-5f2f1d7afaf2. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 878.874594] env[62460]: DEBUG oslo_concurrency.lockutils [req-ae7ec45c-8c02-43bd-b5de-48648b1b557f req-a43aefc6-d189-4dfd-953f-8302a2ac26ea service nova] Acquiring lock "refresh_cache-3e10af70-db52-4d4a-bb92-821a05dcbab6" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.874889] env[62460]: DEBUG oslo_concurrency.lockutils [req-ae7ec45c-8c02-43bd-b5de-48648b1b557f req-a43aefc6-d189-4dfd-953f-8302a2ac26ea service nova] Acquired lock "refresh_cache-3e10af70-db52-4d4a-bb92-821a05dcbab6" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.875231] env[62460]: DEBUG nova.network.neutron [req-ae7ec45c-8c02-43bd-b5de-48648b1b557f req-a43aefc6-d189-4dfd-953f-8302a2ac26ea service nova] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Refreshing network info cache for port 530faf0d-567e-464f-99de-5f2f1d7afaf2 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 878.940388] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa8bf709-4a74-4381-ba6e-93a51a383d76 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.949710] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7dc87e3-1e12-4bc3-8416-4c6af844c371 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.982135] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c0863b1-d597-40c4-9206-7279984240d9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.989698] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7b33979-a0d0-4ad5-8dd3-d192e937fdf9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.002812] env[62460]: DEBUG nova.compute.provider_tree [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 879.075377] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 879.075681] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-686d29c8-bce2-4df2-8fd8-4cada361299c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.084711] env[62460]: DEBUG oslo_vmware.api [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 879.084711] env[62460]: value = "task-1313749" [ 879.084711] env[62460]: _type = "Task" [ 879.084711] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.092816] env[62460]: DEBUG oslo_vmware.api [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313749, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.276785] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d726dbd9-d102-4b1f-957b-cb46a9ba9e28 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Acquiring lock "72e91f8e-0619-464c-b9bc-d6a14be42cb8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.277037] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d726dbd9-d102-4b1f-957b-cb46a9ba9e28 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Lock "72e91f8e-0619-464c-b9bc-d6a14be42cb8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.277279] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d726dbd9-d102-4b1f-957b-cb46a9ba9e28 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Acquiring lock "72e91f8e-0619-464c-b9bc-d6a14be42cb8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.277486] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d726dbd9-d102-4b1f-957b-cb46a9ba9e28 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Lock "72e91f8e-0619-464c-b9bc-d6a14be42cb8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.277715] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d726dbd9-d102-4b1f-957b-cb46a9ba9e28 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Lock "72e91f8e-0619-464c-b9bc-d6a14be42cb8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.280234] env[62460]: INFO nova.compute.manager [None req-d726dbd9-d102-4b1f-957b-cb46a9ba9e28 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Terminating instance [ 879.282191] env[62460]: DEBUG nova.compute.manager [None req-d726dbd9-d102-4b1f-957b-cb46a9ba9e28 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 879.282419] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-d726dbd9-d102-4b1f-957b-cb46a9ba9e28 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 879.283693] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab362887-0f04-492d-938c-f8b8ec157a2f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.021954] env[62460]: DEBUG nova.scheduler.client.report [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 880.025446] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 880.026059] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 880.035870] env[62460]: DEBUG oslo_vmware.api [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313749, 'name': PowerOffVM_Task, 'duration_secs': 0.269826} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.038048] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 880.038295] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 880.038601] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-d726dbd9-d102-4b1f-957b-cb46a9ba9e28 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 880.039846] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92da520d-26f1-4a94-99ce-48c0d0cba99d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.042301] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c816a46d-eae3-4d53-8d29-d31b2d490faa {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.048312] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 880.050164] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0f465dcc-a60d-498f-b82a-c766058e8a71 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.051679] env[62460]: DEBUG oslo_vmware.api [None req-d726dbd9-d102-4b1f-957b-cb46a9ba9e28 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Waiting for the task: (returnval){ [ 880.051679] env[62460]: value = "task-1313751" [ 880.051679] env[62460]: _type = "Task" [ 880.051679] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.064944] env[62460]: DEBUG oslo_vmware.api [None req-d726dbd9-d102-4b1f-957b-cb46a9ba9e28 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313751, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.133526] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 880.133761] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Deleting contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 880.133945] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Deleting the datastore file [datastore2] fd03509c-5471-48bb-8150-d29531dca848 {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 880.134261] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-28534f85-8a32-465a-bf1f-21fd642b53cc {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.141284] env[62460]: DEBUG oslo_vmware.api [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 880.141284] env[62460]: value = "task-1313753" [ 880.141284] env[62460]: _type = "Task" [ 880.141284] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.151805] env[62460]: DEBUG oslo_vmware.api [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313753, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.513787] env[62460]: DEBUG nova.network.neutron [req-ae7ec45c-8c02-43bd-b5de-48648b1b557f req-a43aefc6-d189-4dfd-953f-8302a2ac26ea service nova] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Updated VIF entry in instance network info cache for port 530faf0d-567e-464f-99de-5f2f1d7afaf2. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 880.514202] env[62460]: DEBUG nova.network.neutron [req-ae7ec45c-8c02-43bd-b5de-48648b1b557f req-a43aefc6-d189-4dfd-953f-8302a2ac26ea service nova] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Updating instance_info_cache with network_info: [{"id": "530faf0d-567e-464f-99de-5f2f1d7afaf2", "address": "fa:16:3e:7d:d3:e1", "network": {"id": "51277201-fa18-437f-95a2-e42936d56791", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1816940505-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.187", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "335cfe76a8c44686891e17906c96d158", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "84aee122-f630-43c5-9cc1-3a38d3819c82", "external-id": "nsx-vlan-transportzone-816", "segmentation_id": 816, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap530faf0d-56", "ovs_interfaceid": "530faf0d-567e-464f-99de-5f2f1d7afaf2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 880.527407] env[62460]: DEBUG oslo_concurrency.lockutils [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.790s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.530152] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.196s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.531568] env[62460]: INFO nova.compute.claims [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 880.537266] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 880.537413] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Starting heal instance info cache {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 880.565639] env[62460]: DEBUG oslo_vmware.api [None req-d726dbd9-d102-4b1f-957b-cb46a9ba9e28 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313751, 'name': PowerOffVM_Task, 'duration_secs': 0.19321} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.566308] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-d726dbd9-d102-4b1f-957b-cb46a9ba9e28 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 880.566750] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-d726dbd9-d102-4b1f-957b-cb46a9ba9e28 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 880.567730] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7a98ad02-5ad6-438b-a3e0-4f581be9dadc {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.630026] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-d726dbd9-d102-4b1f-957b-cb46a9ba9e28 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 880.630026] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-d726dbd9-d102-4b1f-957b-cb46a9ba9e28 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Deleting contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 880.630026] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-d726dbd9-d102-4b1f-957b-cb46a9ba9e28 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Deleting the datastore file [datastore2] 72e91f8e-0619-464c-b9bc-d6a14be42cb8 {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 880.630026] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e7b50083-6df8-463c-a78b-c66eb1c7368e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.636431] env[62460]: DEBUG oslo_vmware.api [None req-d726dbd9-d102-4b1f-957b-cb46a9ba9e28 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Waiting for the task: (returnval){ [ 880.636431] env[62460]: value = "task-1313755" [ 880.636431] env[62460]: _type = "Task" [ 880.636431] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.644579] env[62460]: DEBUG oslo_vmware.api [None req-d726dbd9-d102-4b1f-957b-cb46a9ba9e28 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313755, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.652012] env[62460]: DEBUG oslo_vmware.api [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313753, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.433007} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.652258] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 880.652449] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Deleted contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 880.652631] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 881.017151] env[62460]: DEBUG oslo_concurrency.lockutils [req-ae7ec45c-8c02-43bd-b5de-48648b1b557f req-a43aefc6-d189-4dfd-953f-8302a2ac26ea service nova] Releasing lock "refresh_cache-3e10af70-db52-4d4a-bb92-821a05dcbab6" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.035248] env[62460]: DEBUG oslo_concurrency.lockutils [None req-14dad4fb-a665-464f-b2d9-35a62b8afd9a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "d7d180f7-e1a8-46c8-ba8e-ca50dac474cc" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 31.582s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.035675] env[62460]: DEBUG oslo_concurrency.lockutils [None req-02a178f4-1158-4294-a7b0-5612d682103c tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "d7d180f7-e1a8-46c8-ba8e-ca50dac474cc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 11.010s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.035979] env[62460]: DEBUG oslo_concurrency.lockutils [None req-02a178f4-1158-4294-a7b0-5612d682103c tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "d7d180f7-e1a8-46c8-ba8e-ca50dac474cc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.036317] env[62460]: DEBUG oslo_concurrency.lockutils [None req-02a178f4-1158-4294-a7b0-5612d682103c tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "d7d180f7-e1a8-46c8-ba8e-ca50dac474cc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.036395] env[62460]: DEBUG oslo_concurrency.lockutils [None req-02a178f4-1158-4294-a7b0-5612d682103c tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "d7d180f7-e1a8-46c8-ba8e-ca50dac474cc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.039852] env[62460]: INFO nova.compute.manager [None req-02a178f4-1158-4294-a7b0-5612d682103c tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Terminating instance [ 881.042994] env[62460]: DEBUG nova.compute.manager [None req-02a178f4-1158-4294-a7b0-5612d682103c tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 881.042994] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-02a178f4-1158-4294-a7b0-5612d682103c tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 881.042994] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-efc3dc7d-2d86-4da8-9a5e-9a8e7265fe89 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.052727] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32dbd242-b7e3-4f8c-add8-1dedc5293f81 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.081283] env[62460]: WARNING nova.virt.vmwareapi.vmops [None req-02a178f4-1158-4294-a7b0-5612d682103c tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d7d180f7-e1a8-46c8-ba8e-ca50dac474cc could not be found. [ 881.081523] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-02a178f4-1158-4294-a7b0-5612d682103c tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 881.081703] env[62460]: INFO nova.compute.manager [None req-02a178f4-1158-4294-a7b0-5612d682103c tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Took 0.04 seconds to destroy the instance on the hypervisor. [ 881.081957] env[62460]: DEBUG oslo.service.loopingcall [None req-02a178f4-1158-4294-a7b0-5612d682103c tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 881.082201] env[62460]: DEBUG nova.compute.manager [-] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 881.082297] env[62460]: DEBUG nova.network.neutron [-] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 881.146415] env[62460]: DEBUG oslo_vmware.api [None req-d726dbd9-d102-4b1f-957b-cb46a9ba9e28 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313755, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.565528] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Acquiring lock "refresh_cache-7e6ff902-4a04-43d5-9014-38c4ec88efc4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.565528] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Acquired lock "refresh_cache-7e6ff902-4a04-43d5-9014-38c4ec88efc4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.566091] env[62460]: DEBUG nova.network.neutron [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Forcefully refreshing network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 881.650090] env[62460]: DEBUG oslo_vmware.api [None req-d726dbd9-d102-4b1f-957b-cb46a9ba9e28 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Task: {'id': task-1313755, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.652467} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.650381] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-d726dbd9-d102-4b1f-957b-cb46a9ba9e28 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 881.650579] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-d726dbd9-d102-4b1f-957b-cb46a9ba9e28 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Deleted contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 881.650996] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-d726dbd9-d102-4b1f-957b-cb46a9ba9e28 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 881.651079] env[62460]: INFO nova.compute.manager [None req-d726dbd9-d102-4b1f-957b-cb46a9ba9e28 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Took 2.37 seconds to destroy the instance on the hypervisor. [ 881.651286] env[62460]: DEBUG oslo.service.loopingcall [None req-d726dbd9-d102-4b1f-957b-cb46a9ba9e28 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 881.651482] env[62460]: DEBUG nova.compute.manager [-] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 881.652125] env[62460]: DEBUG nova.network.neutron [-] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 881.692351] env[62460]: DEBUG nova.virt.hardware [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 881.692614] env[62460]: DEBUG nova.virt.hardware [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 881.692780] env[62460]: DEBUG nova.virt.hardware [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 881.692969] env[62460]: DEBUG nova.virt.hardware [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 881.693145] env[62460]: DEBUG nova.virt.hardware [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 881.693302] env[62460]: DEBUG nova.virt.hardware [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 881.693519] env[62460]: DEBUG nova.virt.hardware [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 881.693688] env[62460]: DEBUG nova.virt.hardware [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 881.693879] env[62460]: DEBUG nova.virt.hardware [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 881.694197] env[62460]: DEBUG nova.virt.hardware [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 881.694279] env[62460]: DEBUG nova.virt.hardware [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 881.695143] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8619c001-a09b-4db0-803d-d06230c70397 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.706998] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-017f8336-a22a-4cf1-867a-c0d2c9638d0d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.722661] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:dd:7e:4d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '04ccbc7a-cf8d-4ea2-8411-291a1e27df7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f5fba0f8-a3a5-4bcf-856c-52e2f4313154', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 881.730093] env[62460]: DEBUG oslo.service.loopingcall [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 881.731191] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-15b5a62f-433a-41dd-8a74-63304b656114 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Volume attach. Driver type: vmdk {{(pid=62460) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 881.731409] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-15b5a62f-433a-41dd-8a74-63304b656114 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281224', 'volume_id': 'a05c7ae1-7cd9-447e-8453-15a2bcde137e', 'name': 'volume-a05c7ae1-7cd9-447e-8453-15a2bcde137e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5b6e8205-003e-49c0-a73d-be2e032a8272', 'attached_at': '', 'detached_at': '', 'volume_id': 'a05c7ae1-7cd9-447e-8453-15a2bcde137e', 'serial': 'a05c7ae1-7cd9-447e-8453-15a2bcde137e'} {{(pid=62460) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 881.732400] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fd03509c-5471-48bb-8150-d29531dca848] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 881.733100] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71e6f9d8-9d5d-489d-a528-535109f1a04c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.735967] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7756f832-d621-4106-8a10-10e769cf1fc6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.738401] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6fa712ad-30d4-46b7-8f31-700f2d202969 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.768067] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7f8cadd-fadc-4950-b2cf-f34ddb851dad {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.770398] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 881.770398] env[62460]: value = "task-1313756" [ 881.770398] env[62460]: _type = "Task" [ 881.770398] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.771300] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83cda7e0-599c-49a6-9b29-6d77281fe4e7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.796842] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-15b5a62f-433a-41dd-8a74-63304b656114 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Reconfiguring VM instance instance-00000042 to attach disk [datastore1] volume-a05c7ae1-7cd9-447e-8453-15a2bcde137e/volume-a05c7ae1-7cd9-447e-8453-15a2bcde137e.vmdk or device None with type thin {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 881.820622] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-13d7275b-4ac2-40f7-96ce-b1a4301055b9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.837540] env[62460]: DEBUG nova.network.neutron [-] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.840614] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-871d069c-122f-407d-be86-53002c49a292 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.840860] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313756, 'name': CreateVM_Task} progress is 25%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.847861] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f938ce1-7a45-4326-9461-9ebb0b1151f6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.852070] env[62460]: DEBUG oslo_vmware.api [None req-15b5a62f-433a-41dd-8a74-63304b656114 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Waiting for the task: (returnval){ [ 881.852070] env[62460]: value = "task-1313757" [ 881.852070] env[62460]: _type = "Task" [ 881.852070] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.864234] env[62460]: DEBUG nova.compute.provider_tree [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 881.872579] env[62460]: DEBUG oslo_vmware.api [None req-15b5a62f-433a-41dd-8a74-63304b656114 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1313757, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.084049] env[62460]: DEBUG nova.network.neutron [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 882.284993] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313756, 'name': CreateVM_Task, 'duration_secs': 0.291969} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.285179] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fd03509c-5471-48bb-8150-d29531dca848] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 882.285814] env[62460]: DEBUG oslo_concurrency.lockutils [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.285992] env[62460]: DEBUG oslo_concurrency.lockutils [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.286339] env[62460]: DEBUG oslo_concurrency.lockutils [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 882.286586] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-293d463f-436e-48bc-a46e-1bfd15f4bf62 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.290971] env[62460]: DEBUG oslo_vmware.api [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 882.290971] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]5277655d-2172-01c7-03ec-3b1190af91f1" [ 882.290971] env[62460]: _type = "Task" [ 882.290971] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.297862] env[62460]: DEBUG oslo_vmware.api [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5277655d-2172-01c7-03ec-3b1190af91f1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.342520] env[62460]: INFO nova.compute.manager [-] [instance: d7d180f7-e1a8-46c8-ba8e-ca50dac474cc] Took 1.26 seconds to deallocate network for instance. [ 882.361328] env[62460]: DEBUG oslo_vmware.api [None req-15b5a62f-433a-41dd-8a74-63304b656114 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1313757, 'name': ReconfigVM_Task, 'duration_secs': 0.351938} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.361612] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-15b5a62f-433a-41dd-8a74-63304b656114 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Reconfigured VM instance instance-00000042 to attach disk [datastore1] volume-a05c7ae1-7cd9-447e-8453-15a2bcde137e/volume-a05c7ae1-7cd9-447e-8453-15a2bcde137e.vmdk or device None with type thin {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 882.366251] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0824fd6b-8ac1-412b-8ce3-7652443a50d0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.378333] env[62460]: DEBUG nova.network.neutron [-] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 882.384473] env[62460]: DEBUG oslo_vmware.api [None req-15b5a62f-433a-41dd-8a74-63304b656114 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Waiting for the task: (returnval){ [ 882.384473] env[62460]: value = "task-1313758" [ 882.384473] env[62460]: _type = "Task" [ 882.384473] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.392892] env[62460]: DEBUG oslo_vmware.api [None req-15b5a62f-433a-41dd-8a74-63304b656114 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1313758, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.405395] env[62460]: DEBUG nova.scheduler.client.report [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Updated inventory for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with generation 94 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 882.405649] env[62460]: DEBUG nova.compute.provider_tree [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Updating resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 generation from 94 to 95 during operation: update_inventory {{(pid=62460) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 882.405831] env[62460]: DEBUG nova.compute.provider_tree [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 882.670996] env[62460]: DEBUG nova.network.neutron [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 882.801689] env[62460]: DEBUG oslo_vmware.api [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5277655d-2172-01c7-03ec-3b1190af91f1, 'name': SearchDatastore_Task, 'duration_secs': 0.008371} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.802014] env[62460]: DEBUG oslo_concurrency.lockutils [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.802270] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 882.802510] env[62460]: DEBUG oslo_concurrency.lockutils [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.802671] env[62460]: DEBUG oslo_concurrency.lockutils [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.802850] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 882.803131] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2882bd4b-4002-4dd0-be9d-2862f40d30d3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.810979] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 882.811185] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 882.811857] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e9a6b7fb-1780-4383-8bf6-bca26accf8ce {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.816556] env[62460]: DEBUG oslo_vmware.api [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 882.816556] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52ee976e-d395-1f25-83e4-e3f14e38016e" [ 882.816556] env[62460]: _type = "Task" [ 882.816556] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.823436] env[62460]: DEBUG oslo_vmware.api [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52ee976e-d395-1f25-83e4-e3f14e38016e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.881641] env[62460]: INFO nova.compute.manager [-] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Took 1.23 seconds to deallocate network for instance. [ 882.896105] env[62460]: DEBUG oslo_vmware.api [None req-15b5a62f-433a-41dd-8a74-63304b656114 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1313758, 'name': ReconfigVM_Task, 'duration_secs': 0.133662} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.896424] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-15b5a62f-433a-41dd-8a74-63304b656114 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281224', 'volume_id': 'a05c7ae1-7cd9-447e-8453-15a2bcde137e', 'name': 'volume-a05c7ae1-7cd9-447e-8453-15a2bcde137e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5b6e8205-003e-49c0-a73d-be2e032a8272', 'attached_at': '', 'detached_at': '', 'volume_id': 'a05c7ae1-7cd9-447e-8453-15a2bcde137e', 'serial': 'a05c7ae1-7cd9-447e-8453-15a2bcde137e'} {{(pid=62460) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 882.910776] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.380s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.911344] env[62460]: DEBUG nova.compute.manager [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 882.913801] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.562s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.915194] env[62460]: INFO nova.compute.claims [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 883.173574] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Releasing lock "refresh_cache-7e6ff902-4a04-43d5-9014-38c4ec88efc4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.173972] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Updated the network info_cache for instance {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 883.174050] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 883.174210] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 883.174364] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 883.174515] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 883.174663] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 883.174829] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 883.174945] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62460) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 883.175110] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 883.326897] env[62460]: DEBUG oslo_vmware.api [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52ee976e-d395-1f25-83e4-e3f14e38016e, 'name': SearchDatastore_Task, 'duration_secs': 0.01118} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.327631] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c0659a87-3e8b-490e-846e-f385d9bfeb78 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.332408] env[62460]: DEBUG oslo_vmware.api [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 883.332408] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52c0b31e-a446-d710-8e5f-47327e63c02d" [ 883.332408] env[62460]: _type = "Task" [ 883.332408] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.340516] env[62460]: DEBUG oslo_vmware.api [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52c0b31e-a446-d710-8e5f-47327e63c02d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.368167] env[62460]: DEBUG oslo_concurrency.lockutils [None req-02a178f4-1158-4294-a7b0-5612d682103c tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "d7d180f7-e1a8-46c8-ba8e-ca50dac474cc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.332s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.375780] env[62460]: DEBUG nova.compute.manager [req-412ac949-4dd8-49ca-9230-bd73e25130d1 req-7de91f93-e130-4bc7-a987-ac066b4247c9 service nova] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Received event network-vif-deleted-215f9da2-99ec-4af7-8488-b7185454b285 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 883.392393] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d726dbd9-d102-4b1f-957b-cb46a9ba9e28 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.419197] env[62460]: DEBUG nova.compute.utils [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 883.422472] env[62460]: DEBUG nova.compute.manager [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 883.422472] env[62460]: DEBUG nova.network.neutron [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 883.470203] env[62460]: DEBUG nova.policy [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b4c1b6d1942542a7815aee4ec860259a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '44220d464249478fb1373bc6b146ee0f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 883.678391] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.727290] env[62460]: DEBUG nova.network.neutron [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Successfully created port: 97108d62-5bb6-4013-b38a-4ccd99a02da5 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 883.847888] env[62460]: DEBUG oslo_vmware.api [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52c0b31e-a446-d710-8e5f-47327e63c02d, 'name': SearchDatastore_Task, 'duration_secs': 0.01136} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.848240] env[62460]: DEBUG oslo_concurrency.lockutils [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.848519] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] fd03509c-5471-48bb-8150-d29531dca848/fd03509c-5471-48bb-8150-d29531dca848.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 883.848799] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0f6be315-c62d-4492-b3cc-f1583e3ab27f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.856111] env[62460]: DEBUG oslo_vmware.api [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 883.856111] env[62460]: value = "task-1313759" [ 883.856111] env[62460]: _type = "Task" [ 883.856111] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.863886] env[62460]: DEBUG oslo_vmware.api [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313759, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.923615] env[62460]: DEBUG nova.compute.manager [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 883.938443] env[62460]: DEBUG nova.objects.instance [None req-15b5a62f-433a-41dd-8a74-63304b656114 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lazy-loading 'flavor' on Instance uuid 5b6e8205-003e-49c0-a73d-be2e032a8272 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 884.139298] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9c94219-3645-441c-a3e6-c6fa33f24138 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.148498] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3d31c32-a5a4-413d-9a89-9bd75aeaef5c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.188428] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85cc0622-73e5-4e50-8a87-62caf6f79286 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.198316] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc5f85cc-e4a9-4d47-81a7-2757cab640d5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.213912] env[62460]: DEBUG nova.compute.provider_tree [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 884.365327] env[62460]: DEBUG oslo_vmware.api [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313759, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.442066] env[62460]: DEBUG oslo_concurrency.lockutils [None req-15b5a62f-433a-41dd-8a74-63304b656114 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lock "5b6e8205-003e-49c0-a73d-be2e032a8272" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.332s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.528195] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1913ee28-6e57-4da2-bed6-194f53eb10a9 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Acquiring lock "5b6e8205-003e-49c0-a73d-be2e032a8272" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.528688] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1913ee28-6e57-4da2-bed6-194f53eb10a9 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lock "5b6e8205-003e-49c0-a73d-be2e032a8272" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.529021] env[62460]: DEBUG nova.compute.manager [None req-1913ee28-6e57-4da2-bed6-194f53eb10a9 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 884.530437] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1a75eb5-4832-4f73-83a5-b6db84b21f2a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.539257] env[62460]: DEBUG nova.compute.manager [None req-1913ee28-6e57-4da2-bed6-194f53eb10a9 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62460) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 884.539858] env[62460]: DEBUG nova.objects.instance [None req-1913ee28-6e57-4da2-bed6-194f53eb10a9 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lazy-loading 'flavor' on Instance uuid 5b6e8205-003e-49c0-a73d-be2e032a8272 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 884.716980] env[62460]: DEBUG nova.scheduler.client.report [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 884.755886] env[62460]: DEBUG oslo_concurrency.lockutils [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "09155629-51c8-4043-b1ee-6d5036552a67" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.756141] env[62460]: DEBUG oslo_concurrency.lockutils [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "09155629-51c8-4043-b1ee-6d5036552a67" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.867041] env[62460]: DEBUG oslo_vmware.api [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313759, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.611134} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.867041] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] fd03509c-5471-48bb-8150-d29531dca848/fd03509c-5471-48bb-8150-d29531dca848.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 884.867041] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 884.867328] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c018d86c-a32b-4c36-9564-c967c47a283c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.876549] env[62460]: DEBUG oslo_vmware.api [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 884.876549] env[62460]: value = "task-1313760" [ 884.876549] env[62460]: _type = "Task" [ 884.876549] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.887950] env[62460]: DEBUG oslo_vmware.api [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313760, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.939888] env[62460]: DEBUG nova.compute.manager [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 884.964664] env[62460]: DEBUG nova.virt.hardware [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 884.964923] env[62460]: DEBUG nova.virt.hardware [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 884.965103] env[62460]: DEBUG nova.virt.hardware [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 884.965293] env[62460]: DEBUG nova.virt.hardware [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 884.965447] env[62460]: DEBUG nova.virt.hardware [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 884.965601] env[62460]: DEBUG nova.virt.hardware [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 884.965810] env[62460]: DEBUG nova.virt.hardware [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 884.966040] env[62460]: DEBUG nova.virt.hardware [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 884.966252] env[62460]: DEBUG nova.virt.hardware [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 884.966429] env[62460]: DEBUG nova.virt.hardware [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 884.966611] env[62460]: DEBUG nova.virt.hardware [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 884.967497] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4f8f6e9-9d40-4f7a-a91b-edc6b53cc6a7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.975548] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-805964f6-3576-45db-9295-a6cb9875cf4e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.045015] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-1913ee28-6e57-4da2-bed6-194f53eb10a9 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 885.045435] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9c0d01dd-4545-4eda-8d7f-cfcfdcc2b926 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.052294] env[62460]: DEBUG oslo_vmware.api [None req-1913ee28-6e57-4da2-bed6-194f53eb10a9 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Waiting for the task: (returnval){ [ 885.052294] env[62460]: value = "task-1313761" [ 885.052294] env[62460]: _type = "Task" [ 885.052294] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.060313] env[62460]: DEBUG oslo_vmware.api [None req-1913ee28-6e57-4da2-bed6-194f53eb10a9 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1313761, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.221936] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.308s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.222555] env[62460]: DEBUG nova.compute.manager [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 885.225900] env[62460]: DEBUG oslo_concurrency.lockutils [None req-36066889-905f-461f-99f9-b5b6a17671e2 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.717s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.226226] env[62460]: DEBUG nova.objects.instance [None req-36066889-905f-461f-99f9-b5b6a17671e2 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Lazy-loading 'resources' on Instance uuid a02b4be1-91b0-4254-8d60-654885e24f6b {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 885.257813] env[62460]: DEBUG nova.compute.manager [req-a0ff64fb-b1df-4386-bd25-4fb072c5df18 req-e17d8cc1-38ce-482b-b9f3-ad39c1c7aefd service nova] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Received event network-vif-plugged-97108d62-5bb6-4013-b38a-4ccd99a02da5 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 885.258156] env[62460]: DEBUG oslo_concurrency.lockutils [req-a0ff64fb-b1df-4386-bd25-4fb072c5df18 req-e17d8cc1-38ce-482b-b9f3-ad39c1c7aefd service nova] Acquiring lock "0da70720-197c-4483-93e1-019e0f040c4c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.258324] env[62460]: DEBUG oslo_concurrency.lockutils [req-a0ff64fb-b1df-4386-bd25-4fb072c5df18 req-e17d8cc1-38ce-482b-b9f3-ad39c1c7aefd service nova] Lock "0da70720-197c-4483-93e1-019e0f040c4c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.258504] env[62460]: DEBUG oslo_concurrency.lockutils [req-a0ff64fb-b1df-4386-bd25-4fb072c5df18 req-e17d8cc1-38ce-482b-b9f3-ad39c1c7aefd service nova] Lock "0da70720-197c-4483-93e1-019e0f040c4c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.259144] env[62460]: DEBUG nova.compute.manager [req-a0ff64fb-b1df-4386-bd25-4fb072c5df18 req-e17d8cc1-38ce-482b-b9f3-ad39c1c7aefd service nova] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] No waiting events found dispatching network-vif-plugged-97108d62-5bb6-4013-b38a-4ccd99a02da5 {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 885.259503] env[62460]: WARNING nova.compute.manager [req-a0ff64fb-b1df-4386-bd25-4fb072c5df18 req-e17d8cc1-38ce-482b-b9f3-ad39c1c7aefd service nova] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Received unexpected event network-vif-plugged-97108d62-5bb6-4013-b38a-4ccd99a02da5 for instance with vm_state building and task_state spawning. [ 885.260558] env[62460]: DEBUG nova.compute.manager [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 885.386643] env[62460]: DEBUG oslo_vmware.api [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313760, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066793} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.387091] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 885.387928] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1d79dc1-5ea5-4592-ae37-a94cdd83152c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.409705] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Reconfiguring VM instance instance-0000004e to attach disk [datastore2] fd03509c-5471-48bb-8150-d29531dca848/fd03509c-5471-48bb-8150-d29531dca848.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 885.409998] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c325a41c-82de-4400-818e-d604297b2d63 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.429111] env[62460]: DEBUG oslo_vmware.api [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 885.429111] env[62460]: value = "task-1313762" [ 885.429111] env[62460]: _type = "Task" [ 885.429111] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.436386] env[62460]: DEBUG oslo_vmware.api [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313762, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.562788] env[62460]: DEBUG oslo_vmware.api [None req-1913ee28-6e57-4da2-bed6-194f53eb10a9 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1313761, 'name': PowerOffVM_Task, 'duration_secs': 0.211543} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.563080] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-1913ee28-6e57-4da2-bed6-194f53eb10a9 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 885.563272] env[62460]: DEBUG nova.compute.manager [None req-1913ee28-6e57-4da2-bed6-194f53eb10a9 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 885.564034] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1ecee6b-b549-4db6-a75c-ff2ac142df9e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.729269] env[62460]: DEBUG nova.compute.utils [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 885.734028] env[62460]: DEBUG nova.compute.manager [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 885.734028] env[62460]: DEBUG nova.network.neutron [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 885.771429] env[62460]: DEBUG nova.network.neutron [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Successfully updated port: 97108d62-5bb6-4013-b38a-4ccd99a02da5 {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 885.779947] env[62460]: DEBUG nova.policy [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b4c1b6d1942542a7815aee4ec860259a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '44220d464249478fb1373bc6b146ee0f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 885.787033] env[62460]: DEBUG oslo_concurrency.lockutils [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.792844] env[62460]: DEBUG nova.compute.manager [req-88c14fbb-980a-4cee-8aa4-111165900c6a req-24ab796e-27b7-4937-b4c7-db1669d25f44 service nova] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Received event network-changed-97108d62-5bb6-4013-b38a-4ccd99a02da5 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 885.793110] env[62460]: DEBUG nova.compute.manager [req-88c14fbb-980a-4cee-8aa4-111165900c6a req-24ab796e-27b7-4937-b4c7-db1669d25f44 service nova] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Refreshing instance network info cache due to event network-changed-97108d62-5bb6-4013-b38a-4ccd99a02da5. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 885.793535] env[62460]: DEBUG oslo_concurrency.lockutils [req-88c14fbb-980a-4cee-8aa4-111165900c6a req-24ab796e-27b7-4937-b4c7-db1669d25f44 service nova] Acquiring lock "refresh_cache-0da70720-197c-4483-93e1-019e0f040c4c" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.793535] env[62460]: DEBUG oslo_concurrency.lockutils [req-88c14fbb-980a-4cee-8aa4-111165900c6a req-24ab796e-27b7-4937-b4c7-db1669d25f44 service nova] Acquired lock "refresh_cache-0da70720-197c-4483-93e1-019e0f040c4c" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.794043] env[62460]: DEBUG nova.network.neutron [req-88c14fbb-980a-4cee-8aa4-111165900c6a req-24ab796e-27b7-4937-b4c7-db1669d25f44 service nova] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Refreshing network info cache for port 97108d62-5bb6-4013-b38a-4ccd99a02da5 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 885.916031] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dbcd079-e0ab-4fb2-a30d-00e660207d67 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.923841] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e370a5b-dece-4b96-bd98-cc2a3ba2edb3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.957279] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dda3a86e-9450-40ae-8ae4-7f2be55addd6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.962970] env[62460]: DEBUG oslo_vmware.api [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313762, 'name': ReconfigVM_Task, 'duration_secs': 0.265291} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.963634] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Reconfigured VM instance instance-0000004e to attach disk [datastore2] fd03509c-5471-48bb-8150-d29531dca848/fd03509c-5471-48bb-8150-d29531dca848.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 885.964278] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e74e9c28-8034-4cbb-9c8b-f2dae3d20e9c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.969030] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60233937-a46c-4eb0-8453-e170b53a376e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.974403] env[62460]: DEBUG oslo_vmware.api [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 885.974403] env[62460]: value = "task-1313763" [ 885.974403] env[62460]: _type = "Task" [ 885.974403] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.985120] env[62460]: DEBUG nova.compute.provider_tree [None req-36066889-905f-461f-99f9-b5b6a17671e2 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 885.991042] env[62460]: DEBUG oslo_vmware.api [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313763, 'name': Rename_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.073205] env[62460]: DEBUG nova.network.neutron [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] Successfully created port: 3acd012b-bd26-45c6-8457-4a0630dbcbf1 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 886.077751] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1913ee28-6e57-4da2-bed6-194f53eb10a9 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lock "5b6e8205-003e-49c0-a73d-be2e032a8272" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.549s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.235050] env[62460]: DEBUG nova.compute.manager [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 886.274064] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Acquiring lock "refresh_cache-0da70720-197c-4483-93e1-019e0f040c4c" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.327626] env[62460]: DEBUG nova.network.neutron [req-88c14fbb-980a-4cee-8aa4-111165900c6a req-24ab796e-27b7-4937-b4c7-db1669d25f44 service nova] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 886.438808] env[62460]: DEBUG nova.network.neutron [req-88c14fbb-980a-4cee-8aa4-111165900c6a req-24ab796e-27b7-4937-b4c7-db1669d25f44 service nova] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.489945] env[62460]: DEBUG nova.scheduler.client.report [None req-36066889-905f-461f-99f9-b5b6a17671e2 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 886.493043] env[62460]: DEBUG oslo_vmware.api [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313763, 'name': Rename_Task, 'duration_secs': 0.125918} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.493678] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 886.494143] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8c61a148-f92a-44af-b680-2d344f03262f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.500944] env[62460]: DEBUG oslo_vmware.api [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 886.500944] env[62460]: value = "task-1313764" [ 886.500944] env[62460]: _type = "Task" [ 886.500944] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.508989] env[62460]: DEBUG oslo_vmware.api [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313764, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.941439] env[62460]: DEBUG oslo_concurrency.lockutils [req-88c14fbb-980a-4cee-8aa4-111165900c6a req-24ab796e-27b7-4937-b4c7-db1669d25f44 service nova] Releasing lock "refresh_cache-0da70720-197c-4483-93e1-019e0f040c4c" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.941854] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Acquired lock "refresh_cache-0da70720-197c-4483-93e1-019e0f040c4c" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.942041] env[62460]: DEBUG nova.network.neutron [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 886.983281] env[62460]: DEBUG nova.objects.instance [None req-82084be4-94d0-48eb-9e78-0efe5143456b tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lazy-loading 'flavor' on Instance uuid 5b6e8205-003e-49c0-a73d-be2e032a8272 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 886.994683] env[62460]: DEBUG oslo_concurrency.lockutils [None req-36066889-905f-461f-99f9-b5b6a17671e2 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.769s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.997222] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e0664ea2-f133-40e3-b052-aae0e3574fee tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.047s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.997479] env[62460]: DEBUG nova.objects.instance [None req-e0664ea2-f133-40e3-b052-aae0e3574fee tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Lazy-loading 'resources' on Instance uuid 0f5457a0-c918-4407-9ba9-bfeeb949c6a7 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 887.011751] env[62460]: DEBUG oslo_vmware.api [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313764, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.016674] env[62460]: INFO nova.scheduler.client.report [None req-36066889-905f-461f-99f9-b5b6a17671e2 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Deleted allocations for instance a02b4be1-91b0-4254-8d60-654885e24f6b [ 887.242930] env[62460]: DEBUG nova.compute.manager [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 887.266676] env[62460]: DEBUG nova.virt.hardware [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 887.266957] env[62460]: DEBUG nova.virt.hardware [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 887.267149] env[62460]: DEBUG nova.virt.hardware [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 887.267344] env[62460]: DEBUG nova.virt.hardware [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 887.267501] env[62460]: DEBUG nova.virt.hardware [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 887.267655] env[62460]: DEBUG nova.virt.hardware [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 887.267915] env[62460]: DEBUG nova.virt.hardware [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 887.268151] env[62460]: DEBUG nova.virt.hardware [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 887.268338] env[62460]: DEBUG nova.virt.hardware [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 887.268511] env[62460]: DEBUG nova.virt.hardware [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 887.268692] env[62460]: DEBUG nova.virt.hardware [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 887.269581] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e4f1342-707b-4beb-9010-a515aedc4113 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.277790] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8213290-a42b-4dd2-ae68-3e94f5b9c2d8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.478968] env[62460]: DEBUG nova.network.neutron [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 887.490688] env[62460]: DEBUG nova.compute.manager [req-463fb115-23ac-48ca-af18-77b42a955e69 req-0a91c722-e301-4820-b4db-3929330effca service nova] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] Received event network-vif-plugged-3acd012b-bd26-45c6-8457-4a0630dbcbf1 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 887.490919] env[62460]: DEBUG oslo_concurrency.lockutils [req-463fb115-23ac-48ca-af18-77b42a955e69 req-0a91c722-e301-4820-b4db-3929330effca service nova] Acquiring lock "fbe558f1-ce5a-4429-8833-48bdabe91071-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.491155] env[62460]: DEBUG oslo_concurrency.lockutils [req-463fb115-23ac-48ca-af18-77b42a955e69 req-0a91c722-e301-4820-b4db-3929330effca service nova] Lock "fbe558f1-ce5a-4429-8833-48bdabe91071-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.491341] env[62460]: DEBUG oslo_concurrency.lockutils [req-463fb115-23ac-48ca-af18-77b42a955e69 req-0a91c722-e301-4820-b4db-3929330effca service nova] Lock "fbe558f1-ce5a-4429-8833-48bdabe91071-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.491545] env[62460]: DEBUG nova.compute.manager [req-463fb115-23ac-48ca-af18-77b42a955e69 req-0a91c722-e301-4820-b4db-3929330effca service nova] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] No waiting events found dispatching network-vif-plugged-3acd012b-bd26-45c6-8457-4a0630dbcbf1 {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 887.491697] env[62460]: WARNING nova.compute.manager [req-463fb115-23ac-48ca-af18-77b42a955e69 req-0a91c722-e301-4820-b4db-3929330effca service nova] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] Received unexpected event network-vif-plugged-3acd012b-bd26-45c6-8457-4a0630dbcbf1 for instance with vm_state building and task_state spawning. [ 887.492231] env[62460]: DEBUG oslo_concurrency.lockutils [None req-82084be4-94d0-48eb-9e78-0efe5143456b tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Acquiring lock "refresh_cache-5b6e8205-003e-49c0-a73d-be2e032a8272" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.492441] env[62460]: DEBUG oslo_concurrency.lockutils [None req-82084be4-94d0-48eb-9e78-0efe5143456b tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Acquired lock "refresh_cache-5b6e8205-003e-49c0-a73d-be2e032a8272" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.492623] env[62460]: DEBUG nova.network.neutron [None req-82084be4-94d0-48eb-9e78-0efe5143456b tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 887.492801] env[62460]: DEBUG nova.objects.instance [None req-82084be4-94d0-48eb-9e78-0efe5143456b tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lazy-loading 'info_cache' on Instance uuid 5b6e8205-003e-49c0-a73d-be2e032a8272 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 887.514501] env[62460]: DEBUG oslo_vmware.api [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313764, 'name': PowerOnVM_Task, 'duration_secs': 0.97167} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.514501] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 887.514623] env[62460]: DEBUG nova.compute.manager [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 887.515294] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ea81b96-8f7d-438e-a611-339a48e91e2b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.530426] env[62460]: DEBUG oslo_concurrency.lockutils [None req-36066889-905f-461f-99f9-b5b6a17671e2 tempest-ImagesTestJSON-75209480 tempest-ImagesTestJSON-75209480-project-member] Lock "a02b4be1-91b0-4254-8d60-654885e24f6b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.636s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.650671] env[62460]: DEBUG nova.network.neutron [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] Successfully updated port: 3acd012b-bd26-45c6-8457-4a0630dbcbf1 {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 887.671689] env[62460]: DEBUG nova.network.neutron [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Updating instance_info_cache with network_info: [{"id": "97108d62-5bb6-4013-b38a-4ccd99a02da5", "address": "fa:16:3e:68:17:9c", "network": {"id": "1e15b3ec-bd84-41d8-92d0-66d17a474051", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-641826644-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "44220d464249478fb1373bc6b146ee0f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ce62383-8e84-4e26-955b-74c11392f4c9", "external-id": "nsx-vlan-transportzone-215", "segmentation_id": 215, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97108d62-5b", "ovs_interfaceid": "97108d62-5bb6-4013-b38a-4ccd99a02da5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.724371] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d0a12e5-f3d4-4675-8a73-744342dbe266 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.731907] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35a2e76e-c20f-43fa-9280-994b513f276c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.761413] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b636abb-bc69-4735-95f3-0e2dc1d06d5a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.768321] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d6a7243-d1b5-44a2-8916-aa93be5f2921 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.781203] env[62460]: DEBUG nova.compute.provider_tree [None req-e0664ea2-f133-40e3-b052-aae0e3574fee tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 887.998153] env[62460]: DEBUG nova.objects.base [None req-82084be4-94d0-48eb-9e78-0efe5143456b tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Object Instance<5b6e8205-003e-49c0-a73d-be2e032a8272> lazy-loaded attributes: flavor,info_cache {{(pid=62460) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 888.041830] env[62460]: DEBUG oslo_concurrency.lockutils [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.153205] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Acquiring lock "refresh_cache-fbe558f1-ce5a-4429-8833-48bdabe91071" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.153349] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Acquired lock "refresh_cache-fbe558f1-ce5a-4429-8833-48bdabe91071" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.153512] env[62460]: DEBUG nova.network.neutron [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 888.174340] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Releasing lock "refresh_cache-0da70720-197c-4483-93e1-019e0f040c4c" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.175113] env[62460]: DEBUG nova.compute.manager [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Instance network_info: |[{"id": "97108d62-5bb6-4013-b38a-4ccd99a02da5", "address": "fa:16:3e:68:17:9c", "network": {"id": "1e15b3ec-bd84-41d8-92d0-66d17a474051", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-641826644-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "44220d464249478fb1373bc6b146ee0f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ce62383-8e84-4e26-955b-74c11392f4c9", "external-id": "nsx-vlan-transportzone-215", "segmentation_id": 215, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97108d62-5b", "ovs_interfaceid": "97108d62-5bb6-4013-b38a-4ccd99a02da5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 888.175708] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:68:17:9c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2ce62383-8e84-4e26-955b-74c11392f4c9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '97108d62-5bb6-4013-b38a-4ccd99a02da5', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 888.194688] env[62460]: DEBUG oslo.service.loopingcall [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 888.196196] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 888.196662] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-99a92d12-f8c7-4755-91fc-1bbd7de34bef {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.223837] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 888.223837] env[62460]: value = "task-1313765" [ 888.223837] env[62460]: _type = "Task" [ 888.223837] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.233169] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313765, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.286217] env[62460]: DEBUG nova.scheduler.client.report [None req-e0664ea2-f133-40e3-b052-aae0e3574fee tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 888.734167] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313765, 'name': CreateVM_Task, 'duration_secs': 0.30369} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.734347] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 888.735093] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.735276] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.735613] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 888.735885] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce5846b7-cfca-43b2-bc81-410de946a274 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.738302] env[62460]: DEBUG nova.network.neutron [None req-82084be4-94d0-48eb-9e78-0efe5143456b tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Updating instance_info_cache with network_info: [{"id": "3515dec2-bd6c-4f96-aff1-8dd32de6f7d6", "address": "fa:16:3e:53:9c:d5", "network": {"id": "4cb03486-3bea-4c6a-a019-37486f271107", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-2094026403-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.174", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfc48e301fb04231b88ec28a062bda4e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c7d2575f-b92f-44ec-a863-634cb76631a2", "external-id": "nsx-vlan-transportzone-794", "segmentation_id": 794, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3515dec2-bd", "ovs_interfaceid": "3515dec2-bd6c-4f96-aff1-8dd32de6f7d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.742855] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Waiting for the task: (returnval){ [ 888.742855] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]527d130a-8780-f691-b064-a6dfca3909a6" [ 888.742855] env[62460]: _type = "Task" [ 888.742855] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.751296] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]527d130a-8780-f691-b064-a6dfca3909a6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.776086] env[62460]: DEBUG nova.network.neutron [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 888.790334] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e0664ea2-f133-40e3-b052-aae0e3574fee tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.793s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.792687] env[62460]: DEBUG oslo_concurrency.lockutils [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.372s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.794346] env[62460]: INFO nova.compute.claims [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 888.839900] env[62460]: INFO nova.scheduler.client.report [None req-e0664ea2-f133-40e3-b052-aae0e3574fee tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Deleted allocations for instance 0f5457a0-c918-4407-9ba9-bfeeb949c6a7 [ 888.995010] env[62460]: DEBUG oslo_concurrency.lockutils [None req-df5627bc-3771-4024-8583-4cb37ab37ea8 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "fd03509c-5471-48bb-8150-d29531dca848" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.995010] env[62460]: DEBUG oslo_concurrency.lockutils [None req-df5627bc-3771-4024-8583-4cb37ab37ea8 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "fd03509c-5471-48bb-8150-d29531dca848" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.995010] env[62460]: DEBUG oslo_concurrency.lockutils [None req-df5627bc-3771-4024-8583-4cb37ab37ea8 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "fd03509c-5471-48bb-8150-d29531dca848-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.995010] env[62460]: DEBUG oslo_concurrency.lockutils [None req-df5627bc-3771-4024-8583-4cb37ab37ea8 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "fd03509c-5471-48bb-8150-d29531dca848-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.995010] env[62460]: DEBUG oslo_concurrency.lockutils [None req-df5627bc-3771-4024-8583-4cb37ab37ea8 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "fd03509c-5471-48bb-8150-d29531dca848-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.998544] env[62460]: INFO nova.compute.manager [None req-df5627bc-3771-4024-8583-4cb37ab37ea8 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Terminating instance [ 889.001482] env[62460]: DEBUG nova.compute.manager [None req-df5627bc-3771-4024-8583-4cb37ab37ea8 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 889.001692] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-df5627bc-3771-4024-8583-4cb37ab37ea8 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 889.003420] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-234aef4f-0e36-4ca1-a667-84f466a04915 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.015284] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-df5627bc-3771-4024-8583-4cb37ab37ea8 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 889.015667] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cb677fbf-4a94-4688-9a07-0f9bb14c8f71 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.021440] env[62460]: DEBUG oslo_vmware.api [None req-df5627bc-3771-4024-8583-4cb37ab37ea8 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 889.021440] env[62460]: value = "task-1313766" [ 889.021440] env[62460]: _type = "Task" [ 889.021440] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.030210] env[62460]: DEBUG oslo_vmware.api [None req-df5627bc-3771-4024-8583-4cb37ab37ea8 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313766, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.077693] env[62460]: DEBUG nova.network.neutron [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] Updating instance_info_cache with network_info: [{"id": "3acd012b-bd26-45c6-8457-4a0630dbcbf1", "address": "fa:16:3e:17:4f:ef", "network": {"id": "1e15b3ec-bd84-41d8-92d0-66d17a474051", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-641826644-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "44220d464249478fb1373bc6b146ee0f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ce62383-8e84-4e26-955b-74c11392f4c9", "external-id": "nsx-vlan-transportzone-215", "segmentation_id": 215, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3acd012b-bd", "ovs_interfaceid": "3acd012b-bd26-45c6-8457-4a0630dbcbf1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.242386] env[62460]: DEBUG oslo_concurrency.lockutils [None req-82084be4-94d0-48eb-9e78-0efe5143456b tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Releasing lock "refresh_cache-5b6e8205-003e-49c0-a73d-be2e032a8272" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.253843] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]527d130a-8780-f691-b064-a6dfca3909a6, 'name': SearchDatastore_Task, 'duration_secs': 0.011214} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.254194] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.254436] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 889.254956] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 889.254956] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.255125] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 889.255369] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-582a3b57-fd4a-4f48-a4fd-1abd820196aa {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.264766] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 889.264766] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 889.265468] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4d63daf5-98cf-441c-af6d-48cbb9feb334 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.271574] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Waiting for the task: (returnval){ [ 889.271574] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]523c2dfe-0980-d43b-053d-a509ef15b084" [ 889.271574] env[62460]: _type = "Task" [ 889.271574] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.280224] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]523c2dfe-0980-d43b-053d-a509ef15b084, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.350690] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e0664ea2-f133-40e3-b052-aae0e3574fee tempest-ServersNegativeTestMultiTenantJSON-730821605 tempest-ServersNegativeTestMultiTenantJSON-730821605-project-member] Lock "0f5457a0-c918-4407-9ba9-bfeeb949c6a7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.406s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.519190] env[62460]: DEBUG nova.compute.manager [req-ffba0039-17dd-4ef5-8fc9-fc202bc75859 req-7e8b0437-8d35-49a9-82bf-cee8d544d23d service nova] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] Received event network-changed-3acd012b-bd26-45c6-8457-4a0630dbcbf1 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 889.519405] env[62460]: DEBUG nova.compute.manager [req-ffba0039-17dd-4ef5-8fc9-fc202bc75859 req-7e8b0437-8d35-49a9-82bf-cee8d544d23d service nova] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] Refreshing instance network info cache due to event network-changed-3acd012b-bd26-45c6-8457-4a0630dbcbf1. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 889.519612] env[62460]: DEBUG oslo_concurrency.lockutils [req-ffba0039-17dd-4ef5-8fc9-fc202bc75859 req-7e8b0437-8d35-49a9-82bf-cee8d544d23d service nova] Acquiring lock "refresh_cache-fbe558f1-ce5a-4429-8833-48bdabe91071" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 889.531721] env[62460]: DEBUG oslo_vmware.api [None req-df5627bc-3771-4024-8583-4cb37ab37ea8 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313766, 'name': PowerOffVM_Task, 'duration_secs': 0.233028} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.531987] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-df5627bc-3771-4024-8583-4cb37ab37ea8 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 889.532184] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-df5627bc-3771-4024-8583-4cb37ab37ea8 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 889.532441] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9477bac6-a626-412c-affe-a1c2907ecd0e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.580636] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Releasing lock "refresh_cache-fbe558f1-ce5a-4429-8833-48bdabe91071" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.581038] env[62460]: DEBUG nova.compute.manager [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] Instance network_info: |[{"id": "3acd012b-bd26-45c6-8457-4a0630dbcbf1", "address": "fa:16:3e:17:4f:ef", "network": {"id": "1e15b3ec-bd84-41d8-92d0-66d17a474051", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-641826644-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "44220d464249478fb1373bc6b146ee0f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ce62383-8e84-4e26-955b-74c11392f4c9", "external-id": "nsx-vlan-transportzone-215", "segmentation_id": 215, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3acd012b-bd", "ovs_interfaceid": "3acd012b-bd26-45c6-8457-4a0630dbcbf1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 889.581384] env[62460]: DEBUG oslo_concurrency.lockutils [req-ffba0039-17dd-4ef5-8fc9-fc202bc75859 req-7e8b0437-8d35-49a9-82bf-cee8d544d23d service nova] Acquired lock "refresh_cache-fbe558f1-ce5a-4429-8833-48bdabe91071" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.581572] env[62460]: DEBUG nova.network.neutron [req-ffba0039-17dd-4ef5-8fc9-fc202bc75859 req-7e8b0437-8d35-49a9-82bf-cee8d544d23d service nova] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] Refreshing network info cache for port 3acd012b-bd26-45c6-8457-4a0630dbcbf1 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 889.582917] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:17:4f:ef', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2ce62383-8e84-4e26-955b-74c11392f4c9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3acd012b-bd26-45c6-8457-4a0630dbcbf1', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 889.594579] env[62460]: DEBUG oslo.service.loopingcall [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 889.598908] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 889.599458] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-df5627bc-3771-4024-8583-4cb37ab37ea8 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 889.599676] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-df5627bc-3771-4024-8583-4cb37ab37ea8 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Deleting contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 889.599859] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-df5627bc-3771-4024-8583-4cb37ab37ea8 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Deleting the datastore file [datastore2] fd03509c-5471-48bb-8150-d29531dca848 {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 889.600113] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3b215071-c4e6-4e26-ade4-5f2c39a42009 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.615310] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-603aac58-00d8-4ec4-bd45-5a97f890debc {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.622698] env[62460]: DEBUG oslo_vmware.api [None req-df5627bc-3771-4024-8583-4cb37ab37ea8 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 889.622698] env[62460]: value = "task-1313769" [ 889.622698] env[62460]: _type = "Task" [ 889.622698] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.623670] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 889.623670] env[62460]: value = "task-1313768" [ 889.623670] env[62460]: _type = "Task" [ 889.623670] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.637055] env[62460]: DEBUG oslo_vmware.api [None req-df5627bc-3771-4024-8583-4cb37ab37ea8 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313769, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.640412] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313768, 'name': CreateVM_Task} progress is 6%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.750092] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-82084be4-94d0-48eb-9e78-0efe5143456b tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 889.750198] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dde00382-ed19-4333-bb5b-7192ca8944ed {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.758798] env[62460]: DEBUG oslo_vmware.api [None req-82084be4-94d0-48eb-9e78-0efe5143456b tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Waiting for the task: (returnval){ [ 889.758798] env[62460]: value = "task-1313770" [ 889.758798] env[62460]: _type = "Task" [ 889.758798] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.772647] env[62460]: DEBUG oslo_vmware.api [None req-82084be4-94d0-48eb-9e78-0efe5143456b tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1313770, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.785779] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "6b084ed4-e8f0-4063-bf2a-252740640753" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.786143] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "6b084ed4-e8f0-4063-bf2a-252740640753" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.787541] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]523c2dfe-0980-d43b-053d-a509ef15b084, 'name': SearchDatastore_Task, 'duration_secs': 0.008457} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.788721] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c7a8785e-cbe9-428f-a8d8-68599f70f948 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.794410] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Waiting for the task: (returnval){ [ 889.794410] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52df17d7-b9fb-f499-cd56-7a5e00b78963" [ 889.794410] env[62460]: _type = "Task" [ 889.794410] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.809527] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52df17d7-b9fb-f499-cd56-7a5e00b78963, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.877850] env[62460]: DEBUG nova.network.neutron [req-ffba0039-17dd-4ef5-8fc9-fc202bc75859 req-7e8b0437-8d35-49a9-82bf-cee8d544d23d service nova] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] Updated VIF entry in instance network info cache for port 3acd012b-bd26-45c6-8457-4a0630dbcbf1. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 889.878362] env[62460]: DEBUG nova.network.neutron [req-ffba0039-17dd-4ef5-8fc9-fc202bc75859 req-7e8b0437-8d35-49a9-82bf-cee8d544d23d service nova] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] Updating instance_info_cache with network_info: [{"id": "3acd012b-bd26-45c6-8457-4a0630dbcbf1", "address": "fa:16:3e:17:4f:ef", "network": {"id": "1e15b3ec-bd84-41d8-92d0-66d17a474051", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-641826644-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "44220d464249478fb1373bc6b146ee0f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ce62383-8e84-4e26-955b-74c11392f4c9", "external-id": "nsx-vlan-transportzone-215", "segmentation_id": 215, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3acd012b-bd", "ovs_interfaceid": "3acd012b-bd26-45c6-8457-4a0630dbcbf1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.026302] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c70daea3-e99b-4222-a468-4472788ec492 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.034447] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3389189c-5a2d-431f-84e5-855ced74323b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.066099] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a22b5771-75af-4f56-a187-29e3f2f6ff21 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.073592] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2901921-d718-448b-b632-02565f9cbad2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.087112] env[62460]: DEBUG nova.compute.provider_tree [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 890.137685] env[62460]: DEBUG oslo_vmware.api [None req-df5627bc-3771-4024-8583-4cb37ab37ea8 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313769, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.179445} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.141219] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-df5627bc-3771-4024-8583-4cb37ab37ea8 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 890.141444] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-df5627bc-3771-4024-8583-4cb37ab37ea8 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Deleted contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 890.141630] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-df5627bc-3771-4024-8583-4cb37ab37ea8 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 890.141818] env[62460]: INFO nova.compute.manager [None req-df5627bc-3771-4024-8583-4cb37ab37ea8 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Took 1.14 seconds to destroy the instance on the hypervisor. [ 890.142112] env[62460]: DEBUG oslo.service.loopingcall [None req-df5627bc-3771-4024-8583-4cb37ab37ea8 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 890.142312] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313768, 'name': CreateVM_Task, 'duration_secs': 0.296706} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.142475] env[62460]: DEBUG nova.compute.manager [-] [instance: fd03509c-5471-48bb-8150-d29531dca848] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 890.142568] env[62460]: DEBUG nova.network.neutron [-] [instance: fd03509c-5471-48bb-8150-d29531dca848] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 890.144288] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 890.144842] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.145012] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.145344] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 890.145921] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d1655b04-94a1-4ae5-8ade-160369f86b5d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.151742] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Waiting for the task: (returnval){ [ 890.151742] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52ce67e6-9b36-17da-6f48-86e8e58d479f" [ 890.151742] env[62460]: _type = "Task" [ 890.151742] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.161993] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52ce67e6-9b36-17da-6f48-86e8e58d479f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.271138] env[62460]: DEBUG oslo_vmware.api [None req-82084be4-94d0-48eb-9e78-0efe5143456b tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1313770, 'name': PowerOnVM_Task, 'duration_secs': 0.41777} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.271600] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-82084be4-94d0-48eb-9e78-0efe5143456b tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 890.271743] env[62460]: DEBUG nova.compute.manager [None req-82084be4-94d0-48eb-9e78-0efe5143456b tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 890.272464] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6e11366-36a0-4f8b-902c-82f26a55fa9f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.288975] env[62460]: DEBUG nova.compute.manager [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 890.306815] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52df17d7-b9fb-f499-cd56-7a5e00b78963, 'name': SearchDatastore_Task, 'duration_secs': 0.010298} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.307176] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.307440] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] 0da70720-197c-4483-93e1-019e0f040c4c/0da70720-197c-4483-93e1-019e0f040c4c.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 890.307732] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0caf65c7-3af9-41e9-824b-fe48451eebfd {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.315071] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Waiting for the task: (returnval){ [ 890.315071] env[62460]: value = "task-1313771" [ 890.315071] env[62460]: _type = "Task" [ 890.315071] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.324155] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313771, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.384882] env[62460]: DEBUG oslo_concurrency.lockutils [req-ffba0039-17dd-4ef5-8fc9-fc202bc75859 req-7e8b0437-8d35-49a9-82bf-cee8d544d23d service nova] Releasing lock "refresh_cache-fbe558f1-ce5a-4429-8833-48bdabe91071" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.590165] env[62460]: DEBUG nova.scheduler.client.report [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 890.665332] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52ce67e6-9b36-17da-6f48-86e8e58d479f, 'name': SearchDatastore_Task, 'duration_secs': 0.011225} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.665702] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.665955] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 890.666446] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.666659] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.666921] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 890.667240] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9c9c6e18-c48a-4326-b0aa-cf0bf7bb4f7a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.677327] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 890.677327] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 890.677327] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-03df63ea-88b4-4b08-8592-c1d972381478 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.682745] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Waiting for the task: (returnval){ [ 890.682745] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]522800b7-997a-ff80-3156-18972797ab1f" [ 890.682745] env[62460]: _type = "Task" [ 890.682745] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.690727] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]522800b7-997a-ff80-3156-18972797ab1f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.826399] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.838519] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313771, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.003893] env[62460]: DEBUG nova.network.neutron [-] [instance: fd03509c-5471-48bb-8150-d29531dca848] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.096285] env[62460]: DEBUG oslo_concurrency.lockutils [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.304s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.097598] env[62460]: DEBUG nova.compute.manager [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 891.099876] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.309s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.102033] env[62460]: INFO nova.compute.claims [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 891.195427] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]522800b7-997a-ff80-3156-18972797ab1f, 'name': SearchDatastore_Task, 'duration_secs': 0.012276} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.196695] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7100bf22-0de2-4262-a9d3-df9fabf71d12 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.202934] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Waiting for the task: (returnval){ [ 891.202934] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52736734-93e4-0249-f477-a77b3b87d4d1" [ 891.202934] env[62460]: _type = "Task" [ 891.202934] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.214204] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52736734-93e4-0249-f477-a77b3b87d4d1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.331488] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313771, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.778523} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.331719] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] 0da70720-197c-4483-93e1-019e0f040c4c/0da70720-197c-4483-93e1-019e0f040c4c.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 891.332042] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 891.332301] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7e142ffb-1a71-4ff4-ad90-f160c4f50192 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.339782] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Waiting for the task: (returnval){ [ 891.339782] env[62460]: value = "task-1313772" [ 891.339782] env[62460]: _type = "Task" [ 891.339782] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.348017] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313772, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.508323] env[62460]: INFO nova.compute.manager [-] [instance: fd03509c-5471-48bb-8150-d29531dca848] Took 1.37 seconds to deallocate network for instance. [ 891.558326] env[62460]: DEBUG nova.compute.manager [req-af65c75f-48bd-4b0b-b796-6f9cdb559a22 req-e87e22b2-7933-4fb4-9c8b-00b7ec4e95fb service nova] [instance: fd03509c-5471-48bb-8150-d29531dca848] Received event network-vif-deleted-f5fba0f8-a3a5-4bcf-856c-52e2f4313154 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 891.607133] env[62460]: DEBUG nova.compute.utils [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 891.610484] env[62460]: DEBUG nova.compute.manager [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 891.610702] env[62460]: DEBUG nova.network.neutron [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 891.710484] env[62460]: DEBUG nova.policy [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5f496fe14b64478d9248d11403b4011b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6432a0219df949d09641a8384ee75785', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 891.715663] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52736734-93e4-0249-f477-a77b3b87d4d1, 'name': SearchDatastore_Task, 'duration_secs': 0.011172} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.715968] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.716288] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] fbe558f1-ce5a-4429-8833-48bdabe91071/fbe558f1-ce5a-4429-8833-48bdabe91071.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 891.716589] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8009baee-4d6a-49ae-b8b4-d64638c05112 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.725643] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Waiting for the task: (returnval){ [ 891.725643] env[62460]: value = "task-1313773" [ 891.725643] env[62460]: _type = "Task" [ 891.725643] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.736218] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313773, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.854970] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313772, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.288992} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.855366] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 891.856775] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf009f5d-bd8b-4c42-a31b-117bd2cd48f8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.884563] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Reconfiguring VM instance instance-00000050 to attach disk [datastore2] 0da70720-197c-4483-93e1-019e0f040c4c/0da70720-197c-4483-93e1-019e0f040c4c.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 891.885398] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3525bc5e-703b-488a-90a1-146eb3eebcd5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.907275] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Waiting for the task: (returnval){ [ 891.907275] env[62460]: value = "task-1313774" [ 891.907275] env[62460]: _type = "Task" [ 891.907275] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.916251] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313774, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.015834] env[62460]: DEBUG oslo_concurrency.lockutils [None req-df5627bc-3771-4024-8583-4cb37ab37ea8 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.113785] env[62460]: DEBUG nova.compute.manager [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 892.193530] env[62460]: DEBUG nova.network.neutron [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Successfully created port: 797761fb-cacf-47d3-8617-691a88101883 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 892.242557] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313773, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.364678] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14d90d33-0269-4a72-8d69-ad746ee86316 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.378277] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5739ff2-0a14-4223-8124-9040c83f07e3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.422420] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eb50799-3e33-4164-82d4-fde5eb1d8723 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.433072] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313774, 'name': ReconfigVM_Task, 'duration_secs': 0.349195} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.436837] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Reconfigured VM instance instance-00000050 to attach disk [datastore2] 0da70720-197c-4483-93e1-019e0f040c4c/0da70720-197c-4483-93e1-019e0f040c4c.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 892.438571] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-74e4c5de-602a-4f1c-bdba-20d1cb642a3a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.441757] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd863cb6-7af1-481e-9bb7-10d0e05e978d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.461052] env[62460]: DEBUG nova.compute.provider_tree [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 892.469024] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Waiting for the task: (returnval){ [ 892.469024] env[62460]: value = "task-1313775" [ 892.469024] env[62460]: _type = "Task" [ 892.469024] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.479110] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313775, 'name': Rename_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.742797] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313773, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.846129} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.742797] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] fbe558f1-ce5a-4429-8833-48bdabe91071/fbe558f1-ce5a-4429-8833-48bdabe91071.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 892.743557] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 892.743557] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1ffa2905-3951-4b01-9e41-dd9ebb78b571 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.751448] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Waiting for the task: (returnval){ [ 892.751448] env[62460]: value = "task-1313776" [ 892.751448] env[62460]: _type = "Task" [ 892.751448] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.762263] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313776, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.970021] env[62460]: DEBUG nova.scheduler.client.report [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 892.984146] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313775, 'name': Rename_Task, 'duration_secs': 0.188131} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.984146] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 892.984295] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7cf8189b-f528-4eb7-b1c2-9ec0d7a6b319 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.992892] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Waiting for the task: (returnval){ [ 892.992892] env[62460]: value = "task-1313777" [ 892.992892] env[62460]: _type = "Task" [ 892.992892] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.002975] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313777, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.126663] env[62460]: DEBUG nova.compute.manager [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 893.183701] env[62460]: DEBUG nova.virt.hardware [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 893.183701] env[62460]: DEBUG nova.virt.hardware [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 893.183701] env[62460]: DEBUG nova.virt.hardware [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 893.183701] env[62460]: DEBUG nova.virt.hardware [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 893.184241] env[62460]: DEBUG nova.virt.hardware [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 893.185946] env[62460]: DEBUG nova.virt.hardware [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 893.185946] env[62460]: DEBUG nova.virt.hardware [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 893.186134] env[62460]: DEBUG nova.virt.hardware [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 893.186468] env[62460]: DEBUG nova.virt.hardware [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 893.186653] env[62460]: DEBUG nova.virt.hardware [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 893.186897] env[62460]: DEBUG nova.virt.hardware [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 893.190334] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8135a3f2-b55a-4c75-a680-ad64863ab641 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.197257] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4655ce69-551e-4213-bcb8-7bcacb238d16 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.262480] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313776, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063525} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.262830] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 893.263724] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32b013a0-2c38-425f-a33a-5a0f39a6b714 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.290237] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] Reconfiguring VM instance instance-00000051 to attach disk [datastore2] fbe558f1-ce5a-4429-8833-48bdabe91071/fbe558f1-ce5a-4429-8833-48bdabe91071.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 893.290576] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cc55767a-343f-4b31-acf9-d7b263e9e0cf {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.309865] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Waiting for the task: (returnval){ [ 893.309865] env[62460]: value = "task-1313778" [ 893.309865] env[62460]: _type = "Task" [ 893.309865] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.320123] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313778, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.478808] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.379s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.479629] env[62460]: DEBUG nova.compute.manager [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 893.486052] env[62460]: DEBUG oslo_concurrency.lockutils [None req-f9e6de72-4edb-4918-b758-21cb807f86bc tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.193s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.486052] env[62460]: DEBUG nova.objects.instance [None req-f9e6de72-4edb-4918-b758-21cb807f86bc tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lazy-loading 'resources' on Instance uuid 7e6ff902-4a04-43d5-9014-38c4ec88efc4 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 893.511725] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313777, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.827747] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313778, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.988975] env[62460]: DEBUG nova.compute.utils [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 893.995020] env[62460]: DEBUG nova.compute.manager [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 893.995020] env[62460]: DEBUG nova.network.neutron [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 894.009539] env[62460]: DEBUG nova.compute.manager [req-9699c36a-f8f7-4e7e-8573-2a3e09a1f9f8 req-12d04279-12b9-47b0-b826-daee6b3cbc65 service nova] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Received event network-vif-plugged-797761fb-cacf-47d3-8617-691a88101883 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 894.010250] env[62460]: DEBUG oslo_concurrency.lockutils [req-9699c36a-f8f7-4e7e-8573-2a3e09a1f9f8 req-12d04279-12b9-47b0-b826-daee6b3cbc65 service nova] Acquiring lock "8bcaa672-97c8-466a-812c-5723e08d0c52-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.010250] env[62460]: DEBUG oslo_concurrency.lockutils [req-9699c36a-f8f7-4e7e-8573-2a3e09a1f9f8 req-12d04279-12b9-47b0-b826-daee6b3cbc65 service nova] Lock "8bcaa672-97c8-466a-812c-5723e08d0c52-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.010250] env[62460]: DEBUG oslo_concurrency.lockutils [req-9699c36a-f8f7-4e7e-8573-2a3e09a1f9f8 req-12d04279-12b9-47b0-b826-daee6b3cbc65 service nova] Lock "8bcaa672-97c8-466a-812c-5723e08d0c52-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.010977] env[62460]: DEBUG nova.compute.manager [req-9699c36a-f8f7-4e7e-8573-2a3e09a1f9f8 req-12d04279-12b9-47b0-b826-daee6b3cbc65 service nova] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] No waiting events found dispatching network-vif-plugged-797761fb-cacf-47d3-8617-691a88101883 {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 894.010977] env[62460]: WARNING nova.compute.manager [req-9699c36a-f8f7-4e7e-8573-2a3e09a1f9f8 req-12d04279-12b9-47b0-b826-daee6b3cbc65 service nova] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Received unexpected event network-vif-plugged-797761fb-cacf-47d3-8617-691a88101883 for instance with vm_state building and task_state spawning. [ 894.017530] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313777, 'name': PowerOnVM_Task, 'duration_secs': 0.559151} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.017798] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 894.018073] env[62460]: INFO nova.compute.manager [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Took 9.08 seconds to spawn the instance on the hypervisor. [ 894.018561] env[62460]: DEBUG nova.compute.manager [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 894.019282] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b5fe1c8-bfb9-408e-9cb3-dbfb5ae2c642 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.074344] env[62460]: DEBUG nova.policy [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '632d7ca8cb4243f990d741a3f70f422a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '10ac54db1ef54e249a077bbb3dde4242', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 894.129230] env[62460]: DEBUG nova.network.neutron [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Successfully updated port: 797761fb-cacf-47d3-8617-691a88101883 {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 894.229035] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0317bd39-44c0-4404-aeaf-4875f2fd4de0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.242076] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-487be56b-a915-4cef-ab7e-3efe88e5c72a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.280707] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e4a3dfd-77e2-4e40-95bd-a38f61b2d800 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.286450] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d668dfd-4ffa-4eb7-933b-b674ef63fdb6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.301410] env[62460]: DEBUG nova.compute.provider_tree [None req-f9e6de72-4edb-4918-b758-21cb807f86bc tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 894.320459] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313778, 'name': ReconfigVM_Task, 'duration_secs': 0.711016} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.320736] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] Reconfigured VM instance instance-00000051 to attach disk [datastore2] fbe558f1-ce5a-4429-8833-48bdabe91071/fbe558f1-ce5a-4429-8833-48bdabe91071.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 894.321364] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-00b39d76-535b-4c03-ba6b-4d4ba5630d37 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.331019] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Waiting for the task: (returnval){ [ 894.331019] env[62460]: value = "task-1313779" [ 894.331019] env[62460]: _type = "Task" [ 894.331019] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.336250] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313779, 'name': Rename_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.358290] env[62460]: DEBUG nova.network.neutron [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] Successfully created port: 5102799a-f429-4aa9-aad4-e476aacddf8b {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 894.498476] env[62460]: DEBUG nova.compute.manager [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 894.551214] env[62460]: INFO nova.compute.manager [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Took 24.23 seconds to build instance. [ 894.634363] env[62460]: DEBUG oslo_concurrency.lockutils [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Acquiring lock "refresh_cache-8bcaa672-97c8-466a-812c-5723e08d0c52" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 894.634540] env[62460]: DEBUG oslo_concurrency.lockutils [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Acquired lock "refresh_cache-8bcaa672-97c8-466a-812c-5723e08d0c52" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.634659] env[62460]: DEBUG nova.network.neutron [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 894.808029] env[62460]: DEBUG nova.scheduler.client.report [None req-f9e6de72-4edb-4918-b758-21cb807f86bc tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 894.837129] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313779, 'name': Rename_Task, 'duration_secs': 0.262612} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.837454] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 894.837974] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fbe1ae17-a0db-43a4-869e-f5218bb52b71 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.845530] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Waiting for the task: (returnval){ [ 894.845530] env[62460]: value = "task-1313780" [ 894.845530] env[62460]: _type = "Task" [ 894.845530] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.854612] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313780, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.054213] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Lock "0da70720-197c-4483-93e1-019e0f040c4c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.750s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.104293] env[62460]: DEBUG oslo_concurrency.lockutils [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Acquiring lock "bf28fc98-d9a9-4678-a107-ace2b6503353" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.104578] env[62460]: DEBUG oslo_concurrency.lockutils [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Lock "bf28fc98-d9a9-4678-a107-ace2b6503353" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.169347] env[62460]: DEBUG nova.network.neutron [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 895.309297] env[62460]: DEBUG nova.network.neutron [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Updating instance_info_cache with network_info: [{"id": "797761fb-cacf-47d3-8617-691a88101883", "address": "fa:16:3e:cf:be:95", "network": {"id": "ee4dac34-8b7b-435c-8f6b-ded5d68e65b6", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1133527409-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6432a0219df949d09641a8384ee75785", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92fe29b3-0907-453d-aabb-5559c4bd7c0f", "external-id": "nsx-vlan-transportzone-482", "segmentation_id": 482, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap797761fb-ca", "ovs_interfaceid": "797761fb-cacf-47d3-8617-691a88101883", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 895.312838] env[62460]: DEBUG oslo_concurrency.lockutils [None req-f9e6de72-4edb-4918-b758-21cb807f86bc tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.829s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.315478] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d726dbd9-d102-4b1f-957b-cb46a9ba9e28 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.923s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.315739] env[62460]: DEBUG nova.objects.instance [None req-d726dbd9-d102-4b1f-957b-cb46a9ba9e28 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Lazy-loading 'resources' on Instance uuid 72e91f8e-0619-464c-b9bc-d6a14be42cb8 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 895.334246] env[62460]: INFO nova.scheduler.client.report [None req-f9e6de72-4edb-4918-b758-21cb807f86bc tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Deleted allocations for instance 7e6ff902-4a04-43d5-9014-38c4ec88efc4 [ 895.358092] env[62460]: DEBUG oslo_vmware.api [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313780, 'name': PowerOnVM_Task, 'duration_secs': 0.503807} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.358370] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 895.358578] env[62460]: INFO nova.compute.manager [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] Took 8.12 seconds to spawn the instance on the hypervisor. [ 895.358778] env[62460]: DEBUG nova.compute.manager [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 895.359590] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82391536-8fa5-4f08-b344-ebf0c99bbe3e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.511074] env[62460]: DEBUG nova.compute.manager [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 895.536533] env[62460]: DEBUG nova.virt.hardware [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 895.536794] env[62460]: DEBUG nova.virt.hardware [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 895.536961] env[62460]: DEBUG nova.virt.hardware [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 895.537261] env[62460]: DEBUG nova.virt.hardware [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 895.537433] env[62460]: DEBUG nova.virt.hardware [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 895.537589] env[62460]: DEBUG nova.virt.hardware [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 895.537803] env[62460]: DEBUG nova.virt.hardware [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 895.537972] env[62460]: DEBUG nova.virt.hardware [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 895.538185] env[62460]: DEBUG nova.virt.hardware [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 895.538376] env[62460]: DEBUG nova.virt.hardware [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 895.538886] env[62460]: DEBUG nova.virt.hardware [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 895.540051] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-214a7111-9b00-435a-9386-7bd9821c875e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.547889] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e292f2a-1bd0-4791-80cd-b2f6e6191d11 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.607024] env[62460]: DEBUG nova.compute.manager [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 895.794564] env[62460]: DEBUG nova.compute.manager [req-bb6ae3c1-56c8-4b6a-b8c6-567dca6b0057 req-8951acf2-558a-4090-8405-5f8023de3c60 service nova] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] Received event network-vif-plugged-5102799a-f429-4aa9-aad4-e476aacddf8b {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 895.794847] env[62460]: DEBUG oslo_concurrency.lockutils [req-bb6ae3c1-56c8-4b6a-b8c6-567dca6b0057 req-8951acf2-558a-4090-8405-5f8023de3c60 service nova] Acquiring lock "e90e8c3a-d244-4d04-997e-c15ab1cb2009-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.795101] env[62460]: DEBUG oslo_concurrency.lockutils [req-bb6ae3c1-56c8-4b6a-b8c6-567dca6b0057 req-8951acf2-558a-4090-8405-5f8023de3c60 service nova] Lock "e90e8c3a-d244-4d04-997e-c15ab1cb2009-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.795313] env[62460]: DEBUG oslo_concurrency.lockutils [req-bb6ae3c1-56c8-4b6a-b8c6-567dca6b0057 req-8951acf2-558a-4090-8405-5f8023de3c60 service nova] Lock "e90e8c3a-d244-4d04-997e-c15ab1cb2009-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.795500] env[62460]: DEBUG nova.compute.manager [req-bb6ae3c1-56c8-4b6a-b8c6-567dca6b0057 req-8951acf2-558a-4090-8405-5f8023de3c60 service nova] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] No waiting events found dispatching network-vif-plugged-5102799a-f429-4aa9-aad4-e476aacddf8b {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 895.795681] env[62460]: WARNING nova.compute.manager [req-bb6ae3c1-56c8-4b6a-b8c6-567dca6b0057 req-8951acf2-558a-4090-8405-5f8023de3c60 service nova] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] Received unexpected event network-vif-plugged-5102799a-f429-4aa9-aad4-e476aacddf8b for instance with vm_state building and task_state spawning. [ 895.811448] env[62460]: DEBUG oslo_concurrency.lockutils [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Releasing lock "refresh_cache-8bcaa672-97c8-466a-812c-5723e08d0c52" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.811757] env[62460]: DEBUG nova.compute.manager [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Instance network_info: |[{"id": "797761fb-cacf-47d3-8617-691a88101883", "address": "fa:16:3e:cf:be:95", "network": {"id": "ee4dac34-8b7b-435c-8f6b-ded5d68e65b6", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1133527409-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6432a0219df949d09641a8384ee75785", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92fe29b3-0907-453d-aabb-5559c4bd7c0f", "external-id": "nsx-vlan-transportzone-482", "segmentation_id": 482, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap797761fb-ca", "ovs_interfaceid": "797761fb-cacf-47d3-8617-691a88101883", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 895.812182] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cf:be:95', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '92fe29b3-0907-453d-aabb-5559c4bd7c0f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '797761fb-cacf-47d3-8617-691a88101883', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 895.820258] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Creating folder: Project (6432a0219df949d09641a8384ee75785). Parent ref: group-v281134. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 895.822937] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4c7691ac-b3e0-4006-9ef9-b497ba20a119 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.834450] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Created folder: Project (6432a0219df949d09641a8384ee75785) in parent group-v281134. [ 895.834658] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Creating folder: Instances. Parent ref: group-v281228. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 895.834941] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8f00638f-e056-47cc-8fa1-57c21135891e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.846664] env[62460]: DEBUG oslo_concurrency.lockutils [None req-f9e6de72-4edb-4918-b758-21cb807f86bc tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "7e6ff902-4a04-43d5-9014-38c4ec88efc4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.907s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.847062] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Created folder: Instances in parent group-v281228. [ 895.847318] env[62460]: DEBUG oslo.service.loopingcall [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 895.847510] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 895.850815] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-aa72019e-5501-4778-ac1c-e53445f3fdd0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.876648] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 895.876648] env[62460]: value = "task-1313783" [ 895.876648] env[62460]: _type = "Task" [ 895.876648] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.884454] env[62460]: INFO nova.compute.manager [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] Took 25.55 seconds to build instance. [ 895.892561] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313783, 'name': CreateVM_Task} progress is 6%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.901373] env[62460]: DEBUG nova.network.neutron [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] Successfully updated port: 5102799a-f429-4aa9-aad4-e476aacddf8b {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 896.045820] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ed901ef-8da1-4882-bccc-70b1c114008a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.053604] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d38fbc5-1dc2-4c91-8d7f-2090a654a46f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.085778] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fce4180a-a9eb-4453-b302-7e75dccf8e7e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.094154] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12a41c71-37a7-4463-8709-a875b1a6c9d5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.108823] env[62460]: DEBUG nova.compute.provider_tree [None req-d726dbd9-d102-4b1f-957b-cb46a9ba9e28 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 896.126467] env[62460]: DEBUG oslo_concurrency.lockutils [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.137690] env[62460]: DEBUG nova.compute.manager [req-c0f27973-45f1-4a86-b92f-e8abdb5d9767 req-903a3248-bfaf-4264-890d-41f7cf025910 service nova] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Received event network-changed-797761fb-cacf-47d3-8617-691a88101883 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 896.137911] env[62460]: DEBUG nova.compute.manager [req-c0f27973-45f1-4a86-b92f-e8abdb5d9767 req-903a3248-bfaf-4264-890d-41f7cf025910 service nova] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Refreshing instance network info cache due to event network-changed-797761fb-cacf-47d3-8617-691a88101883. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 896.138148] env[62460]: DEBUG oslo_concurrency.lockutils [req-c0f27973-45f1-4a86-b92f-e8abdb5d9767 req-903a3248-bfaf-4264-890d-41f7cf025910 service nova] Acquiring lock "refresh_cache-8bcaa672-97c8-466a-812c-5723e08d0c52" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.138405] env[62460]: DEBUG oslo_concurrency.lockutils [req-c0f27973-45f1-4a86-b92f-e8abdb5d9767 req-903a3248-bfaf-4264-890d-41f7cf025910 service nova] Acquired lock "refresh_cache-8bcaa672-97c8-466a-812c-5723e08d0c52" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.138554] env[62460]: DEBUG nova.network.neutron [req-c0f27973-45f1-4a86-b92f-e8abdb5d9767 req-903a3248-bfaf-4264-890d-41f7cf025910 service nova] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Refreshing network info cache for port 797761fb-cacf-47d3-8617-691a88101883 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 896.387296] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5251d800-b266-4085-a845-b69ce7126a7d tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Lock "fbe558f1-ce5a-4429-8833-48bdabe91071" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.075s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.387552] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313783, 'name': CreateVM_Task, 'duration_secs': 0.387737} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.387727] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 896.388505] env[62460]: DEBUG oslo_concurrency.lockutils [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.388679] env[62460]: DEBUG oslo_concurrency.lockutils [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.389048] env[62460]: DEBUG oslo_concurrency.lockutils [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 896.389455] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fd702156-2a79-4f2b-af34-7a971b5e8dda {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.394446] env[62460]: DEBUG oslo_vmware.api [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Waiting for the task: (returnval){ [ 896.394446] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52ce9225-65a8-55d9-1361-326ef10d76a5" [ 896.394446] env[62460]: _type = "Task" [ 896.394446] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.403352] env[62460]: DEBUG oslo_vmware.api [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52ce9225-65a8-55d9-1361-326ef10d76a5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.403855] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "refresh_cache-e90e8c3a-d244-4d04-997e-c15ab1cb2009" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.404015] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquired lock "refresh_cache-e90e8c3a-d244-4d04-997e-c15ab1cb2009" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.404177] env[62460]: DEBUG nova.network.neutron [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 896.614015] env[62460]: DEBUG nova.scheduler.client.report [None req-d726dbd9-d102-4b1f-957b-cb46a9ba9e28 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 896.911328] env[62460]: DEBUG oslo_vmware.api [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52ce9225-65a8-55d9-1361-326ef10d76a5, 'name': SearchDatastore_Task, 'duration_secs': 0.011995} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.913605] env[62460]: DEBUG oslo_concurrency.lockutils [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.913892] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 896.914198] env[62460]: DEBUG oslo_concurrency.lockutils [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.914404] env[62460]: DEBUG oslo_concurrency.lockutils [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.914642] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 896.915237] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dd830240-5753-4eb4-9c73-edce45a31076 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.925289] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 896.925488] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 896.926215] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8668f9a5-6e9e-452d-aa18-f0734d96a189 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.931924] env[62460]: DEBUG oslo_vmware.api [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Waiting for the task: (returnval){ [ 896.931924] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]529573b5-e7d8-d1a8-2ad0-872f16d381d4" [ 896.931924] env[62460]: _type = "Task" [ 896.931924] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.939416] env[62460]: DEBUG oslo_vmware.api [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]529573b5-e7d8-d1a8-2ad0-872f16d381d4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.946785] env[62460]: DEBUG nova.network.neutron [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 896.980501] env[62460]: DEBUG nova.network.neutron [req-c0f27973-45f1-4a86-b92f-e8abdb5d9767 req-903a3248-bfaf-4264-890d-41f7cf025910 service nova] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Updated VIF entry in instance network info cache for port 797761fb-cacf-47d3-8617-691a88101883. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 896.980878] env[62460]: DEBUG nova.network.neutron [req-c0f27973-45f1-4a86-b92f-e8abdb5d9767 req-903a3248-bfaf-4264-890d-41f7cf025910 service nova] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Updating instance_info_cache with network_info: [{"id": "797761fb-cacf-47d3-8617-691a88101883", "address": "fa:16:3e:cf:be:95", "network": {"id": "ee4dac34-8b7b-435c-8f6b-ded5d68e65b6", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1133527409-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6432a0219df949d09641a8384ee75785", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92fe29b3-0907-453d-aabb-5559c4bd7c0f", "external-id": "nsx-vlan-transportzone-482", "segmentation_id": 482, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap797761fb-ca", "ovs_interfaceid": "797761fb-cacf-47d3-8617-691a88101883", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.051993] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b138fff5-d93a-4df7-93dc-d05c5fa66426 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Acquiring lock "0da70720-197c-4483-93e1-019e0f040c4c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.052325] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b138fff5-d93a-4df7-93dc-d05c5fa66426 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Lock "0da70720-197c-4483-93e1-019e0f040c4c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.052584] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b138fff5-d93a-4df7-93dc-d05c5fa66426 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Acquiring lock "0da70720-197c-4483-93e1-019e0f040c4c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.053350] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b138fff5-d93a-4df7-93dc-d05c5fa66426 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Lock "0da70720-197c-4483-93e1-019e0f040c4c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.053582] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b138fff5-d93a-4df7-93dc-d05c5fa66426 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Lock "0da70720-197c-4483-93e1-019e0f040c4c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.056235] env[62460]: INFO nova.compute.manager [None req-b138fff5-d93a-4df7-93dc-d05c5fa66426 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Terminating instance [ 897.058365] env[62460]: DEBUG nova.compute.manager [None req-b138fff5-d93a-4df7-93dc-d05c5fa66426 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 897.058612] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-b138fff5-d93a-4df7-93dc-d05c5fa66426 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 897.059559] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-781a8679-2157-4ef1-b063-7071fbc5deb9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.069139] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-b138fff5-d93a-4df7-93dc-d05c5fa66426 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 897.070537] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bcf16a5a-8ea2-4bfe-800b-7a5a76060d0e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.077790] env[62460]: DEBUG oslo_vmware.api [None req-b138fff5-d93a-4df7-93dc-d05c5fa66426 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Waiting for the task: (returnval){ [ 897.077790] env[62460]: value = "task-1313784" [ 897.077790] env[62460]: _type = "Task" [ 897.077790] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.088899] env[62460]: DEBUG oslo_vmware.api [None req-b138fff5-d93a-4df7-93dc-d05c5fa66426 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313784, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.103831] env[62460]: DEBUG nova.network.neutron [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] Updating instance_info_cache with network_info: [{"id": "5102799a-f429-4aa9-aad4-e476aacddf8b", "address": "fa:16:3e:b0:ca:8f", "network": {"id": "c8b3d824-c8f0-4f7b-a91f-1987e04a5f09", "bridge": "br-int", "label": "tempest-ServersTestJSON-1769082812-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10ac54db1ef54e249a077bbb3dde4242", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "255460d5-71d4-4bfd-87f1-acc10085db7f", "external-id": "nsx-vlan-transportzone-152", "segmentation_id": 152, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5102799a-f4", "ovs_interfaceid": "5102799a-f429-4aa9-aad4-e476aacddf8b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.119347] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d726dbd9-d102-4b1f-957b-cb46a9ba9e28 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.804s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.122488] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 13.444s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.122667] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.122829] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62460) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 897.123231] env[62460]: DEBUG oslo_concurrency.lockutils [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.336s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.124724] env[62460]: INFO nova.compute.claims [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 897.128395] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a804a375-a4e7-4118-8997-76c65de0dcd0 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Acquiring lock "fbe558f1-ce5a-4429-8833-48bdabe91071" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.128724] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a804a375-a4e7-4118-8997-76c65de0dcd0 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Lock "fbe558f1-ce5a-4429-8833-48bdabe91071" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.129014] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a804a375-a4e7-4118-8997-76c65de0dcd0 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Acquiring lock "fbe558f1-ce5a-4429-8833-48bdabe91071-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.129297] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a804a375-a4e7-4118-8997-76c65de0dcd0 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Lock "fbe558f1-ce5a-4429-8833-48bdabe91071-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.129651] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a804a375-a4e7-4118-8997-76c65de0dcd0 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Lock "fbe558f1-ce5a-4429-8833-48bdabe91071-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.131972] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db709025-c840-455a-b302-ad8952cb7ab3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.136888] env[62460]: INFO nova.compute.manager [None req-a804a375-a4e7-4118-8997-76c65de0dcd0 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] Terminating instance [ 897.139695] env[62460]: DEBUG nova.compute.manager [None req-a804a375-a4e7-4118-8997-76c65de0dcd0 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 897.139912] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a804a375-a4e7-4118-8997-76c65de0dcd0 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 897.140746] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9d7b200-4c25-41ff-b596-4cf5c2c38274 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.146966] env[62460]: INFO nova.scheduler.client.report [None req-d726dbd9-d102-4b1f-957b-cb46a9ba9e28 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Deleted allocations for instance 72e91f8e-0619-464c-b9bc-d6a14be42cb8 [ 897.149597] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4eada1bb-cc18-401f-8605-370b9caf802c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.159856] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-a804a375-a4e7-4118-8997-76c65de0dcd0 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 897.161451] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c0f19d18-b0e1-4a43-b4df-aaf3e18cc2c1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.174806] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cfca81b-6d13-4ee2-a91d-c9a6c37f396c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.180086] env[62460]: DEBUG oslo_vmware.api [None req-a804a375-a4e7-4118-8997-76c65de0dcd0 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Waiting for the task: (returnval){ [ 897.180086] env[62460]: value = "task-1313785" [ 897.180086] env[62460]: _type = "Task" [ 897.180086] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.190028] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9654268c-28e5-4c7f-98ee-efe7725ce954 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.195025] env[62460]: DEBUG oslo_vmware.api [None req-a804a375-a4e7-4118-8997-76c65de0dcd0 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313785, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.227394] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181031MB free_disk=135GB free_vcpus=48 pci_devices=None {{(pid=62460) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 897.230955] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.443749] env[62460]: DEBUG oslo_vmware.api [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]529573b5-e7d8-d1a8-2ad0-872f16d381d4, 'name': SearchDatastore_Task, 'duration_secs': 0.037858} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.444522] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-269a8796-410f-46d4-bf57-2035ecbc32d5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.449980] env[62460]: DEBUG oslo_vmware.api [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Waiting for the task: (returnval){ [ 897.449980] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]523c081c-5f5a-a002-8198-cedc4c4e186c" [ 897.449980] env[62460]: _type = "Task" [ 897.449980] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.457428] env[62460]: DEBUG oslo_vmware.api [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]523c081c-5f5a-a002-8198-cedc4c4e186c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.485410] env[62460]: DEBUG oslo_concurrency.lockutils [req-c0f27973-45f1-4a86-b92f-e8abdb5d9767 req-903a3248-bfaf-4264-890d-41f7cf025910 service nova] Releasing lock "refresh_cache-8bcaa672-97c8-466a-812c-5723e08d0c52" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.587829] env[62460]: DEBUG oslo_vmware.api [None req-b138fff5-d93a-4df7-93dc-d05c5fa66426 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313784, 'name': PowerOffVM_Task, 'duration_secs': 0.251329} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.588180] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-b138fff5-d93a-4df7-93dc-d05c5fa66426 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 897.588381] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-b138fff5-d93a-4df7-93dc-d05c5fa66426 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 897.588859] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-126ea208-9c5c-4d7d-b8be-35846704b069 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.606248] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Releasing lock "refresh_cache-e90e8c3a-d244-4d04-997e-c15ab1cb2009" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.606570] env[62460]: DEBUG nova.compute.manager [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] Instance network_info: |[{"id": "5102799a-f429-4aa9-aad4-e476aacddf8b", "address": "fa:16:3e:b0:ca:8f", "network": {"id": "c8b3d824-c8f0-4f7b-a91f-1987e04a5f09", "bridge": "br-int", "label": "tempest-ServersTestJSON-1769082812-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10ac54db1ef54e249a077bbb3dde4242", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "255460d5-71d4-4bfd-87f1-acc10085db7f", "external-id": "nsx-vlan-transportzone-152", "segmentation_id": 152, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5102799a-f4", "ovs_interfaceid": "5102799a-f429-4aa9-aad4-e476aacddf8b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 897.606978] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b0:ca:8f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '255460d5-71d4-4bfd-87f1-acc10085db7f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5102799a-f429-4aa9-aad4-e476aacddf8b', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 897.620089] env[62460]: DEBUG oslo.service.loopingcall [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 897.620362] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 897.620607] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d3cfc82b-3fc2-4946-858e-353d70734ad9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.643388] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 897.643388] env[62460]: value = "task-1313787" [ 897.643388] env[62460]: _type = "Task" [ 897.643388] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.644429] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-b138fff5-d93a-4df7-93dc-d05c5fa66426 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 897.644668] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-b138fff5-d93a-4df7-93dc-d05c5fa66426 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Deleting contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 897.644886] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-b138fff5-d93a-4df7-93dc-d05c5fa66426 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Deleting the datastore file [datastore2] 0da70720-197c-4483-93e1-019e0f040c4c {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 897.648489] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4ad5060d-e227-4c75-af77-d2452d3f3dc7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.656412] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313787, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.658143] env[62460]: DEBUG oslo_vmware.api [None req-b138fff5-d93a-4df7-93dc-d05c5fa66426 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Waiting for the task: (returnval){ [ 897.658143] env[62460]: value = "task-1313788" [ 897.658143] env[62460]: _type = "Task" [ 897.658143] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.664156] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d726dbd9-d102-4b1f-957b-cb46a9ba9e28 tempest-ListServerFiltersTestJSON-1962117767 tempest-ListServerFiltersTestJSON-1962117767-project-member] Lock "72e91f8e-0619-464c-b9bc-d6a14be42cb8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.387s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.669519] env[62460]: DEBUG oslo_vmware.api [None req-b138fff5-d93a-4df7-93dc-d05c5fa66426 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313788, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.687024] env[62460]: DEBUG oslo_vmware.api [None req-a804a375-a4e7-4118-8997-76c65de0dcd0 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313785, 'name': PowerOffVM_Task, 'duration_secs': 0.211995} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.688077] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-a804a375-a4e7-4118-8997-76c65de0dcd0 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 897.688077] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a804a375-a4e7-4118-8997-76c65de0dcd0 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 897.688782] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6e0cb215-5f0b-41ef-ac37-37ce152be674 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.761460] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a804a375-a4e7-4118-8997-76c65de0dcd0 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 897.761734] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a804a375-a4e7-4118-8997-76c65de0dcd0 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] Deleting contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 897.761976] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-a804a375-a4e7-4118-8997-76c65de0dcd0 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Deleting the datastore file [datastore2] fbe558f1-ce5a-4429-8833-48bdabe91071 {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 897.762305] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-95c8270a-6d77-46ac-8f07-6829c73a79e9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.770789] env[62460]: DEBUG oslo_vmware.api [None req-a804a375-a4e7-4118-8997-76c65de0dcd0 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Waiting for the task: (returnval){ [ 897.770789] env[62460]: value = "task-1313790" [ 897.770789] env[62460]: _type = "Task" [ 897.770789] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.779665] env[62460]: DEBUG oslo_vmware.api [None req-a804a375-a4e7-4118-8997-76c65de0dcd0 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313790, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.819741] env[62460]: DEBUG nova.compute.manager [req-be0be694-679f-46f9-b36b-edce9099fe71 req-79372c9f-a5b7-4f82-b324-0b8ecf7b4bd7 service nova] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] Received event network-changed-5102799a-f429-4aa9-aad4-e476aacddf8b {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 897.819961] env[62460]: DEBUG nova.compute.manager [req-be0be694-679f-46f9-b36b-edce9099fe71 req-79372c9f-a5b7-4f82-b324-0b8ecf7b4bd7 service nova] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] Refreshing instance network info cache due to event network-changed-5102799a-f429-4aa9-aad4-e476aacddf8b. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 897.820202] env[62460]: DEBUG oslo_concurrency.lockutils [req-be0be694-679f-46f9-b36b-edce9099fe71 req-79372c9f-a5b7-4f82-b324-0b8ecf7b4bd7 service nova] Acquiring lock "refresh_cache-e90e8c3a-d244-4d04-997e-c15ab1cb2009" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.820360] env[62460]: DEBUG oslo_concurrency.lockutils [req-be0be694-679f-46f9-b36b-edce9099fe71 req-79372c9f-a5b7-4f82-b324-0b8ecf7b4bd7 service nova] Acquired lock "refresh_cache-e90e8c3a-d244-4d04-997e-c15ab1cb2009" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.820532] env[62460]: DEBUG nova.network.neutron [req-be0be694-679f-46f9-b36b-edce9099fe71 req-79372c9f-a5b7-4f82-b324-0b8ecf7b4bd7 service nova] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] Refreshing network info cache for port 5102799a-f429-4aa9-aad4-e476aacddf8b {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 897.961954] env[62460]: DEBUG oslo_vmware.api [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]523c081c-5f5a-a002-8198-cedc4c4e186c, 'name': SearchDatastore_Task, 'duration_secs': 0.016255} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.962465] env[62460]: DEBUG oslo_concurrency.lockutils [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.962868] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 8bcaa672-97c8-466a-812c-5723e08d0c52/8bcaa672-97c8-466a-812c-5723e08d0c52.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 897.963271] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-82b35077-033f-4639-a726-c3b5a22ca6ca {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.970286] env[62460]: DEBUG oslo_vmware.api [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Waiting for the task: (returnval){ [ 897.970286] env[62460]: value = "task-1313791" [ 897.970286] env[62460]: _type = "Task" [ 897.970286] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.978077] env[62460]: DEBUG oslo_vmware.api [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Task: {'id': task-1313791, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.119423] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "abd4501b-3ad8-4103-921d-5b80fc2f313f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.119423] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "abd4501b-3ad8-4103-921d-5b80fc2f313f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.156735] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313787, 'name': CreateVM_Task, 'duration_secs': 0.327894} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.157276] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 898.158187] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.158520] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.158955] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 898.159399] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c4d48587-ae65-4be3-ac24-6cbe97fa0883 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.170601] env[62460]: DEBUG oslo_vmware.api [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for the task: (returnval){ [ 898.170601] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52ce889b-8383-b727-7c9a-bbe0094b48fb" [ 898.170601] env[62460]: _type = "Task" [ 898.170601] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.177742] env[62460]: DEBUG oslo_vmware.api [None req-b138fff5-d93a-4df7-93dc-d05c5fa66426 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313788, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.176864} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.181934] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-b138fff5-d93a-4df7-93dc-d05c5fa66426 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 898.183100] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-b138fff5-d93a-4df7-93dc-d05c5fa66426 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Deleted contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 898.183100] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-b138fff5-d93a-4df7-93dc-d05c5fa66426 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 898.183100] env[62460]: INFO nova.compute.manager [None req-b138fff5-d93a-4df7-93dc-d05c5fa66426 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Took 1.12 seconds to destroy the instance on the hypervisor. [ 898.183462] env[62460]: DEBUG oslo.service.loopingcall [None req-b138fff5-d93a-4df7-93dc-d05c5fa66426 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 898.188272] env[62460]: DEBUG nova.compute.manager [-] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 898.188487] env[62460]: DEBUG nova.network.neutron [-] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 898.200223] env[62460]: DEBUG oslo_vmware.api [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52ce889b-8383-b727-7c9a-bbe0094b48fb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.284161] env[62460]: DEBUG oslo_vmware.api [None req-a804a375-a4e7-4118-8997-76c65de0dcd0 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Task: {'id': task-1313790, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.194071} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.284524] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-a804a375-a4e7-4118-8997-76c65de0dcd0 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 898.284703] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a804a375-a4e7-4118-8997-76c65de0dcd0 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] Deleted contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 898.284879] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a804a375-a4e7-4118-8997-76c65de0dcd0 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 898.285093] env[62460]: INFO nova.compute.manager [None req-a804a375-a4e7-4118-8997-76c65de0dcd0 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] Took 1.15 seconds to destroy the instance on the hypervisor. [ 898.285367] env[62460]: DEBUG oslo.service.loopingcall [None req-a804a375-a4e7-4118-8997-76c65de0dcd0 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 898.285601] env[62460]: DEBUG nova.compute.manager [-] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 898.285701] env[62460]: DEBUG nova.network.neutron [-] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 898.384920] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-696ec314-65d8-4254-bcaf-7a3ba822acde {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.394926] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07c42096-ecff-4442-8975-391f51bcc4d2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.425211] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-033aa29d-1222-4997-9f32-56d85d58cd72 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.433240] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f134eef-8e10-4dcf-b1dc-c5f29a7cdc60 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.448825] env[62460]: DEBUG nova.compute.provider_tree [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 898.480808] env[62460]: DEBUG oslo_vmware.api [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Task: {'id': task-1313791, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.499225} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.481041] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 8bcaa672-97c8-466a-812c-5723e08d0c52/8bcaa672-97c8-466a-812c-5723e08d0c52.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 898.481333] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 898.481624] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-310aad43-4399-4b11-863a-6df7ce51a364 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.487870] env[62460]: DEBUG oslo_vmware.api [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Waiting for the task: (returnval){ [ 898.487870] env[62460]: value = "task-1313792" [ 898.487870] env[62460]: _type = "Task" [ 898.487870] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.496017] env[62460]: DEBUG oslo_vmware.api [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Task: {'id': task-1313792, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.620746] env[62460]: DEBUG nova.compute.manager [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 898.674647] env[62460]: DEBUG nova.compute.manager [req-155ef686-c92f-40da-9256-9e8aa8082da0 req-3734ed3a-6c66-41d8-b8f8-106f113a0c07 service nova] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Received event network-vif-deleted-97108d62-5bb6-4013-b38a-4ccd99a02da5 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 898.674870] env[62460]: INFO nova.compute.manager [req-155ef686-c92f-40da-9256-9e8aa8082da0 req-3734ed3a-6c66-41d8-b8f8-106f113a0c07 service nova] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Neutron deleted interface 97108d62-5bb6-4013-b38a-4ccd99a02da5; detaching it from the instance and deleting it from the info cache [ 898.675152] env[62460]: DEBUG nova.network.neutron [req-155ef686-c92f-40da-9256-9e8aa8082da0 req-3734ed3a-6c66-41d8-b8f8-106f113a0c07 service nova] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 898.676757] env[62460]: DEBUG nova.network.neutron [req-be0be694-679f-46f9-b36b-edce9099fe71 req-79372c9f-a5b7-4f82-b324-0b8ecf7b4bd7 service nova] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] Updated VIF entry in instance network info cache for port 5102799a-f429-4aa9-aad4-e476aacddf8b. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 898.677163] env[62460]: DEBUG nova.network.neutron [req-be0be694-679f-46f9-b36b-edce9099fe71 req-79372c9f-a5b7-4f82-b324-0b8ecf7b4bd7 service nova] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] Updating instance_info_cache with network_info: [{"id": "5102799a-f429-4aa9-aad4-e476aacddf8b", "address": "fa:16:3e:b0:ca:8f", "network": {"id": "c8b3d824-c8f0-4f7b-a91f-1987e04a5f09", "bridge": "br-int", "label": "tempest-ServersTestJSON-1769082812-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10ac54db1ef54e249a077bbb3dde4242", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "255460d5-71d4-4bfd-87f1-acc10085db7f", "external-id": "nsx-vlan-transportzone-152", "segmentation_id": 152, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5102799a-f4", "ovs_interfaceid": "5102799a-f429-4aa9-aad4-e476aacddf8b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 898.688889] env[62460]: DEBUG oslo_vmware.api [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52ce889b-8383-b727-7c9a-bbe0094b48fb, 'name': SearchDatastore_Task, 'duration_secs': 0.063971} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.689200] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.689426] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 898.689658] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.689809] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.689997] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 898.690704] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5fbf15a0-50b6-4f7b-b36e-d17f1e54096c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.699381] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 898.699599] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 898.700495] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eb63dc67-c9e7-47e7-ba38-975679225a54 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.705885] env[62460]: DEBUG oslo_vmware.api [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for the task: (returnval){ [ 898.705885] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]529bd6bd-1325-b51c-39ae-a92125c3c6ad" [ 898.705885] env[62460]: _type = "Task" [ 898.705885] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.713794] env[62460]: DEBUG oslo_vmware.api [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]529bd6bd-1325-b51c-39ae-a92125c3c6ad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.954871] env[62460]: DEBUG nova.scheduler.client.report [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 898.997433] env[62460]: DEBUG oslo_vmware.api [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Task: {'id': task-1313792, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.094608} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.997783] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 898.998584] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0054507-7ed6-40fe-aba9-63928565fd26 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.020272] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Reconfiguring VM instance instance-00000052 to attach disk [datastore1] 8bcaa672-97c8-466a-812c-5723e08d0c52/8bcaa672-97c8-466a-812c-5723e08d0c52.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 899.020527] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d2a5268e-4867-4569-9343-7b75f80d042e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.040623] env[62460]: DEBUG oslo_vmware.api [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Waiting for the task: (returnval){ [ 899.040623] env[62460]: value = "task-1313793" [ 899.040623] env[62460]: _type = "Task" [ 899.040623] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.050952] env[62460]: DEBUG oslo_vmware.api [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Task: {'id': task-1313793, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.085661] env[62460]: DEBUG nova.network.neutron [-] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.145041] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.151696] env[62460]: DEBUG nova.network.neutron [-] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.180090] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d5046368-a424-4a29-8fb0-9d2e96bed88a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.182575] env[62460]: DEBUG oslo_concurrency.lockutils [req-be0be694-679f-46f9-b36b-edce9099fe71 req-79372c9f-a5b7-4f82-b324-0b8ecf7b4bd7 service nova] Releasing lock "refresh_cache-e90e8c3a-d244-4d04-997e-c15ab1cb2009" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.190085] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51928860-c7a6-49ac-a510-2c73ee701c42 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.217728] env[62460]: DEBUG nova.compute.manager [req-155ef686-c92f-40da-9256-9e8aa8082da0 req-3734ed3a-6c66-41d8-b8f8-106f113a0c07 service nova] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Detach interface failed, port_id=97108d62-5bb6-4013-b38a-4ccd99a02da5, reason: Instance 0da70720-197c-4483-93e1-019e0f040c4c could not be found. {{(pid=62460) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 899.221443] env[62460]: DEBUG oslo_vmware.api [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]529bd6bd-1325-b51c-39ae-a92125c3c6ad, 'name': SearchDatastore_Task, 'duration_secs': 0.01066} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.222343] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-48d3403e-8a03-4743-a7fa-970327497ffd {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.228465] env[62460]: DEBUG oslo_vmware.api [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for the task: (returnval){ [ 899.228465] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52ac780e-9aee-2fa2-2e19-ce236cabbe17" [ 899.228465] env[62460]: _type = "Task" [ 899.228465] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.237049] env[62460]: DEBUG oslo_vmware.api [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52ac780e-9aee-2fa2-2e19-ce236cabbe17, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.460434] env[62460]: DEBUG oslo_concurrency.lockutils [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.337s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.461175] env[62460]: DEBUG nova.compute.manager [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 899.464046] env[62460]: DEBUG oslo_concurrency.lockutils [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 11.422s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.464171] env[62460]: DEBUG nova.objects.instance [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: fd03509c-5471-48bb-8150-d29531dca848] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62460) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 899.551009] env[62460]: DEBUG oslo_vmware.api [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Task: {'id': task-1313793, 'name': ReconfigVM_Task, 'duration_secs': 0.314985} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.551342] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Reconfigured VM instance instance-00000052 to attach disk [datastore1] 8bcaa672-97c8-466a-812c-5723e08d0c52/8bcaa672-97c8-466a-812c-5723e08d0c52.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 899.552015] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a1d4933c-c085-4715-9fce-31f628aee729 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.558548] env[62460]: DEBUG oslo_vmware.api [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Waiting for the task: (returnval){ [ 899.558548] env[62460]: value = "task-1313794" [ 899.558548] env[62460]: _type = "Task" [ 899.558548] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.568283] env[62460]: DEBUG oslo_vmware.api [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Task: {'id': task-1313794, 'name': Rename_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.589064] env[62460]: INFO nova.compute.manager [-] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] Took 1.30 seconds to deallocate network for instance. [ 899.654973] env[62460]: INFO nova.compute.manager [-] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Took 1.47 seconds to deallocate network for instance. [ 899.738275] env[62460]: DEBUG oslo_vmware.api [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52ac780e-9aee-2fa2-2e19-ce236cabbe17, 'name': SearchDatastore_Task, 'duration_secs': 0.016422} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.738562] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.738908] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] e90e8c3a-d244-4d04-997e-c15ab1cb2009/e90e8c3a-d244-4d04-997e-c15ab1cb2009.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 899.739186] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2f2959d6-2e10-4c20-919c-08d885cb7b73 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.744948] env[62460]: DEBUG oslo_vmware.api [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for the task: (returnval){ [ 899.744948] env[62460]: value = "task-1313795" [ 899.744948] env[62460]: _type = "Task" [ 899.744948] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.752262] env[62460]: DEBUG oslo_vmware.api [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313795, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.843828] env[62460]: DEBUG nova.compute.manager [req-ac1a9306-cc59-45a5-805b-d309545a7dee req-4218a2a5-77ac-4a30-9761-ae605d4fdab4 service nova] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] Received event network-vif-deleted-3acd012b-bd26-45c6-8457-4a0630dbcbf1 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 899.968750] env[62460]: DEBUG nova.compute.utils [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 899.973368] env[62460]: DEBUG nova.compute.manager [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 899.973675] env[62460]: DEBUG nova.network.neutron [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 900.018762] env[62460]: DEBUG nova.policy [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5f69b3e3f50248ff8e57054294a04686', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0da056d93bdf40c39d6e82e457727ff6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 900.069662] env[62460]: DEBUG oslo_vmware.api [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Task: {'id': task-1313794, 'name': Rename_Task, 'duration_secs': 0.130594} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.070146] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 900.070433] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3ee7a63f-4214-4276-b500-c35f3e07bda0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.077326] env[62460]: DEBUG oslo_vmware.api [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Waiting for the task: (returnval){ [ 900.077326] env[62460]: value = "task-1313796" [ 900.077326] env[62460]: _type = "Task" [ 900.077326] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.085692] env[62460]: DEBUG oslo_vmware.api [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Task: {'id': task-1313796, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.096959] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a804a375-a4e7-4118-8997-76c65de0dcd0 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.162745] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b138fff5-d93a-4df7-93dc-d05c5fa66426 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.255098] env[62460]: DEBUG oslo_vmware.api [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313795, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.305016] env[62460]: DEBUG nova.network.neutron [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Successfully created port: dbd267c2-b2d8-46ac-8d82-f645dd8560d5 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 900.312586] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquiring lock "896f8ff4-e45f-4403-a727-03ee25e58609" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.312846] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lock "896f8ff4-e45f-4403-a727-03ee25e58609" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.313040] env[62460]: INFO nova.compute.manager [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Shelving [ 900.474487] env[62460]: DEBUG nova.compute.manager [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 900.478643] env[62460]: DEBUG oslo_concurrency.lockutils [None req-fb62b482-6abf-4e13-ad6d-d4ec988d907c tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.015s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.480412] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.655s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.481928] env[62460]: INFO nova.compute.claims [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 900.589390] env[62460]: DEBUG oslo_vmware.api [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Task: {'id': task-1313796, 'name': PowerOnVM_Task, 'duration_secs': 0.442415} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.589390] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 900.589390] env[62460]: INFO nova.compute.manager [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Took 7.46 seconds to spawn the instance on the hypervisor. [ 900.589743] env[62460]: DEBUG nova.compute.manager [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 900.590672] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03f6d26c-1870-4b94-84d6-8e0da6f69343 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.755264] env[62460]: DEBUG oslo_vmware.api [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313795, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.821506] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 900.821834] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-abd9a955-f790-4517-b93e-5b76dd7a8b7a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.828646] env[62460]: DEBUG oslo_vmware.api [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 900.828646] env[62460]: value = "task-1313797" [ 900.828646] env[62460]: _type = "Task" [ 900.828646] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.836356] env[62460]: DEBUG oslo_vmware.api [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313797, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.109544] env[62460]: INFO nova.compute.manager [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Took 27.70 seconds to build instance. [ 901.257288] env[62460]: DEBUG oslo_vmware.api [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313795, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.462148} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.257487] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] e90e8c3a-d244-4d04-997e-c15ab1cb2009/e90e8c3a-d244-4d04-997e-c15ab1cb2009.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 901.257712] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 901.257980] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f39bf094-0275-456f-9d04-c2b703fa70c6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.264219] env[62460]: DEBUG oslo_vmware.api [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for the task: (returnval){ [ 901.264219] env[62460]: value = "task-1313798" [ 901.264219] env[62460]: _type = "Task" [ 901.264219] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.271888] env[62460]: DEBUG oslo_vmware.api [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313798, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.338217] env[62460]: DEBUG oslo_vmware.api [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313797, 'name': PowerOffVM_Task, 'duration_secs': 0.500979} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.338605] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 901.339356] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-176250bd-e8ee-405e-b4f9-84bdc4b9b2df {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.358191] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc6fa805-22d5-4a1a-ba0f-37e0f08226a2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.487277] env[62460]: DEBUG nova.compute.manager [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 901.513555] env[62460]: DEBUG nova.virt.hardware [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 901.513555] env[62460]: DEBUG nova.virt.hardware [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 901.513735] env[62460]: DEBUG nova.virt.hardware [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 901.513903] env[62460]: DEBUG nova.virt.hardware [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 901.514088] env[62460]: DEBUG nova.virt.hardware [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 901.514255] env[62460]: DEBUG nova.virt.hardware [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 901.514530] env[62460]: DEBUG nova.virt.hardware [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 901.514715] env[62460]: DEBUG nova.virt.hardware [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 901.514894] env[62460]: DEBUG nova.virt.hardware [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 901.515084] env[62460]: DEBUG nova.virt.hardware [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 901.515276] env[62460]: DEBUG nova.virt.hardware [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 901.516260] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63ac0db7-f40c-49f9-a404-a84270e7274b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.525842] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae3fe585-0ccb-491d-8ae2-98ea354fda58 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.612066] env[62460]: DEBUG oslo_concurrency.lockutils [None req-59eff1a6-dc66-4b2f-afbf-bf31bce755ad tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Lock "8bcaa672-97c8-466a-812c-5723e08d0c52" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.214s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.655825] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dffb3160-cf81-4b6d-ae47-0026803d119b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.663336] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b63e6a49-4dd5-48cd-8523-632292268171 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.694161] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c595123-3023-4b10-b95f-689c303bd507 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.702411] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53b62719-d12b-44e7-9e8a-b96149b4aa91 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.715985] env[62460]: DEBUG nova.compute.provider_tree [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 901.776466] env[62460]: DEBUG oslo_vmware.api [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313798, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069597} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.776911] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 901.777578] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0367b518-532a-47e6-ba30-a14883a54977 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.803251] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] Reconfiguring VM instance instance-00000053 to attach disk [datastore1] e90e8c3a-d244-4d04-997e-c15ab1cb2009/e90e8c3a-d244-4d04-997e-c15ab1cb2009.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 901.803567] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-972a8c21-15b4-4235-9046-53617b6cf35c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.823512] env[62460]: DEBUG oslo_vmware.api [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for the task: (returnval){ [ 901.823512] env[62460]: value = "task-1313799" [ 901.823512] env[62460]: _type = "Task" [ 901.823512] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.831162] env[62460]: DEBUG oslo_vmware.api [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313799, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.871348] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Creating Snapshot of the VM instance {{(pid=62460) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 901.871348] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-ea58a19b-482c-4800-92b5-f86b384be472 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.885042] env[62460]: DEBUG oslo_vmware.api [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 901.885042] env[62460]: value = "task-1313800" [ 901.885042] env[62460]: _type = "Task" [ 901.885042] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.885042] env[62460]: DEBUG nova.compute.manager [req-5594f3ff-50cf-4020-9f15-b17fbd708bd6 req-342efdf5-d4d9-4d49-957a-0fef6b93f815 service nova] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Received event network-vif-plugged-dbd267c2-b2d8-46ac-8d82-f645dd8560d5 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 901.885042] env[62460]: DEBUG oslo_concurrency.lockutils [req-5594f3ff-50cf-4020-9f15-b17fbd708bd6 req-342efdf5-d4d9-4d49-957a-0fef6b93f815 service nova] Acquiring lock "09155629-51c8-4043-b1ee-6d5036552a67-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.885042] env[62460]: DEBUG oslo_concurrency.lockutils [req-5594f3ff-50cf-4020-9f15-b17fbd708bd6 req-342efdf5-d4d9-4d49-957a-0fef6b93f815 service nova] Lock "09155629-51c8-4043-b1ee-6d5036552a67-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.885042] env[62460]: DEBUG oslo_concurrency.lockutils [req-5594f3ff-50cf-4020-9f15-b17fbd708bd6 req-342efdf5-d4d9-4d49-957a-0fef6b93f815 service nova] Lock "09155629-51c8-4043-b1ee-6d5036552a67-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.885042] env[62460]: DEBUG nova.compute.manager [req-5594f3ff-50cf-4020-9f15-b17fbd708bd6 req-342efdf5-d4d9-4d49-957a-0fef6b93f815 service nova] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] No waiting events found dispatching network-vif-plugged-dbd267c2-b2d8-46ac-8d82-f645dd8560d5 {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 901.885042] env[62460]: WARNING nova.compute.manager [req-5594f3ff-50cf-4020-9f15-b17fbd708bd6 req-342efdf5-d4d9-4d49-957a-0fef6b93f815 service nova] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Received unexpected event network-vif-plugged-dbd267c2-b2d8-46ac-8d82-f645dd8560d5 for instance with vm_state building and task_state spawning. [ 901.893564] env[62460]: DEBUG oslo_vmware.api [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313800, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.014428] env[62460]: DEBUG nova.network.neutron [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Successfully updated port: dbd267c2-b2d8-46ac-8d82-f645dd8560d5 {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 902.094841] env[62460]: DEBUG oslo_concurrency.lockutils [None req-07205f12-9c2f-4768-9b9d-0f27bbfe3d9c tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Acquiring lock "f1921907-f69c-49a3-896c-a4b12b526ffa" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.095214] env[62460]: DEBUG oslo_concurrency.lockutils [None req-07205f12-9c2f-4768-9b9d-0f27bbfe3d9c tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Lock "f1921907-f69c-49a3-896c-a4b12b526ffa" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.095474] env[62460]: DEBUG oslo_concurrency.lockutils [None req-07205f12-9c2f-4768-9b9d-0f27bbfe3d9c tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Acquiring lock "f1921907-f69c-49a3-896c-a4b12b526ffa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.095692] env[62460]: DEBUG oslo_concurrency.lockutils [None req-07205f12-9c2f-4768-9b9d-0f27bbfe3d9c tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Lock "f1921907-f69c-49a3-896c-a4b12b526ffa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.095888] env[62460]: DEBUG oslo_concurrency.lockutils [None req-07205f12-9c2f-4768-9b9d-0f27bbfe3d9c tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Lock "f1921907-f69c-49a3-896c-a4b12b526ffa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.100072] env[62460]: INFO nova.compute.manager [None req-07205f12-9c2f-4768-9b9d-0f27bbfe3d9c tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Terminating instance [ 902.102281] env[62460]: DEBUG nova.compute.manager [None req-07205f12-9c2f-4768-9b9d-0f27bbfe3d9c tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 902.102523] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-07205f12-9c2f-4768-9b9d-0f27bbfe3d9c tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 902.102768] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f451c38d-e9aa-4e93-b0c6-0b91ece64b17 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.109504] env[62460]: DEBUG oslo_vmware.api [None req-07205f12-9c2f-4768-9b9d-0f27bbfe3d9c tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Waiting for the task: (returnval){ [ 902.109504] env[62460]: value = "task-1313801" [ 902.109504] env[62460]: _type = "Task" [ 902.109504] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.117474] env[62460]: DEBUG oslo_vmware.api [None req-07205f12-9c2f-4768-9b9d-0f27bbfe3d9c tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Task: {'id': task-1313801, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.240204] env[62460]: ERROR nova.scheduler.client.report [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [req-1152276d-b0bb-4a18-9f9d-1666f7b4daa7] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 76e8f00d-65cc-4766-bb9d-5006fb51c728. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-1152276d-b0bb-4a18-9f9d-1666f7b4daa7"}]} [ 902.258255] env[62460]: DEBUG nova.scheduler.client.report [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Refreshing inventories for resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 902.274549] env[62460]: DEBUG nova.scheduler.client.report [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Updating ProviderTree inventory for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 902.274549] env[62460]: DEBUG nova.compute.provider_tree [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 902.287563] env[62460]: DEBUG nova.scheduler.client.report [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Refreshing aggregate associations for resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728, aggregates: None {{(pid=62460) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 902.305355] env[62460]: DEBUG nova.scheduler.client.report [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Refreshing trait associations for resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62460) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 902.334571] env[62460]: DEBUG oslo_vmware.api [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313799, 'name': ReconfigVM_Task, 'duration_secs': 0.258133} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.334872] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] Reconfigured VM instance instance-00000053 to attach disk [datastore1] e90e8c3a-d244-4d04-997e-c15ab1cb2009/e90e8c3a-d244-4d04-997e-c15ab1cb2009.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 902.335597] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a5512068-baa6-4172-a4a3-f8cd0865d200 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.342669] env[62460]: DEBUG oslo_vmware.api [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for the task: (returnval){ [ 902.342669] env[62460]: value = "task-1313802" [ 902.342669] env[62460]: _type = "Task" [ 902.342669] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.350263] env[62460]: DEBUG oslo_vmware.api [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313802, 'name': Rename_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.355861] env[62460]: DEBUG oslo_concurrency.lockutils [None req-8f649042-5fd7-4f94-80db-d6a078b9ce5a tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Acquiring lock "8bcaa672-97c8-466a-812c-5723e08d0c52" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.356106] env[62460]: DEBUG oslo_concurrency.lockutils [None req-8f649042-5fd7-4f94-80db-d6a078b9ce5a tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Lock "8bcaa672-97c8-466a-812c-5723e08d0c52" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.356315] env[62460]: DEBUG oslo_concurrency.lockutils [None req-8f649042-5fd7-4f94-80db-d6a078b9ce5a tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Acquiring lock "8bcaa672-97c8-466a-812c-5723e08d0c52-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.356523] env[62460]: DEBUG oslo_concurrency.lockutils [None req-8f649042-5fd7-4f94-80db-d6a078b9ce5a tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Lock "8bcaa672-97c8-466a-812c-5723e08d0c52-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.356732] env[62460]: DEBUG oslo_concurrency.lockutils [None req-8f649042-5fd7-4f94-80db-d6a078b9ce5a tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Lock "8bcaa672-97c8-466a-812c-5723e08d0c52-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.358771] env[62460]: INFO nova.compute.manager [None req-8f649042-5fd7-4f94-80db-d6a078b9ce5a tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Terminating instance [ 902.362490] env[62460]: DEBUG nova.compute.manager [None req-8f649042-5fd7-4f94-80db-d6a078b9ce5a tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 902.362762] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-8f649042-5fd7-4f94-80db-d6a078b9ce5a tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 902.363689] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-202f4e11-ba99-43c6-b6d7-fdeaa6ddbedf {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.370310] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f649042-5fd7-4f94-80db-d6a078b9ce5a tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 902.370559] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6ce94887-86ac-471c-a06b-4c621fce7939 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.378658] env[62460]: DEBUG oslo_vmware.api [None req-8f649042-5fd7-4f94-80db-d6a078b9ce5a tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Waiting for the task: (returnval){ [ 902.378658] env[62460]: value = "task-1313803" [ 902.378658] env[62460]: _type = "Task" [ 902.378658] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.385857] env[62460]: DEBUG oslo_vmware.api [None req-8f649042-5fd7-4f94-80db-d6a078b9ce5a tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Task: {'id': task-1313803, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.397525] env[62460]: DEBUG oslo_vmware.api [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313800, 'name': CreateSnapshot_Task, 'duration_secs': 0.387764} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.397864] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Created Snapshot of the VM instance {{(pid=62460) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 902.398577] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-302932d1-ec19-46a1-a58a-fb0174508bee {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.474071] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dca2976-8331-473d-9315-fdb4db9f6291 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.482800] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b85f9a1-7090-4df9-9465-80aa56a81670 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.514514] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5df4a71d-e0d5-4d1f-94b5-1140febbd06a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.517236] env[62460]: DEBUG oslo_concurrency.lockutils [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "refresh_cache-09155629-51c8-4043-b1ee-6d5036552a67" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.517394] env[62460]: DEBUG oslo_concurrency.lockutils [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquired lock "refresh_cache-09155629-51c8-4043-b1ee-6d5036552a67" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.517556] env[62460]: DEBUG nova.network.neutron [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 902.523241] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a09cd5f7-ad86-41e1-8da4-c5571ce34f6d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.537464] env[62460]: DEBUG nova.compute.provider_tree [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 902.619131] env[62460]: DEBUG oslo_vmware.api [None req-07205f12-9c2f-4768-9b9d-0f27bbfe3d9c tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Task: {'id': task-1313801, 'name': PowerOffVM_Task, 'duration_secs': 0.17863} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.619463] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-07205f12-9c2f-4768-9b9d-0f27bbfe3d9c tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 902.619684] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-07205f12-9c2f-4768-9b9d-0f27bbfe3d9c tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Volume detach. Driver type: vmdk {{(pid=62460) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 902.619881] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-07205f12-9c2f-4768-9b9d-0f27bbfe3d9c tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281168', 'volume_id': '2cf093c7-9278-4f2d-a385-8775acc7dae8', 'name': 'volume-2cf093c7-9278-4f2d-a385-8775acc7dae8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f1921907-f69c-49a3-896c-a4b12b526ffa', 'attached_at': '', 'detached_at': '', 'volume_id': '2cf093c7-9278-4f2d-a385-8775acc7dae8', 'serial': '2cf093c7-9278-4f2d-a385-8775acc7dae8'} {{(pid=62460) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 902.620779] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf4ea591-7d3d-46c2-aab2-3b5ef814a09a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.637783] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70b60527-4247-4d6b-a119-90814f1a6352 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.644144] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3b74266-9b7c-474b-8c4c-b41b28d34797 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.660786] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abbfd2e6-de53-4699-9da3-8284b23ad7e1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.675147] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-07205f12-9c2f-4768-9b9d-0f27bbfe3d9c tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] The volume has not been displaced from its original location: [datastore2] volume-2cf093c7-9278-4f2d-a385-8775acc7dae8/volume-2cf093c7-9278-4f2d-a385-8775acc7dae8.vmdk. No consolidation needed. {{(pid=62460) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 902.680299] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-07205f12-9c2f-4768-9b9d-0f27bbfe3d9c tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Reconfiguring VM instance instance-0000004b to detach disk 2000 {{(pid=62460) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 902.680554] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3f58afef-0572-45f5-b338-fe4f4854829e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.698188] env[62460]: DEBUG oslo_vmware.api [None req-07205f12-9c2f-4768-9b9d-0f27bbfe3d9c tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Waiting for the task: (returnval){ [ 902.698188] env[62460]: value = "task-1313804" [ 902.698188] env[62460]: _type = "Task" [ 902.698188] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.706157] env[62460]: DEBUG oslo_vmware.api [None req-07205f12-9c2f-4768-9b9d-0f27bbfe3d9c tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Task: {'id': task-1313804, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.852555] env[62460]: DEBUG oslo_vmware.api [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313802, 'name': Rename_Task, 'duration_secs': 0.143314} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.852867] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 902.853131] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-be2b4012-a264-43cd-aa99-5babb6b1bed0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.859312] env[62460]: DEBUG oslo_vmware.api [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for the task: (returnval){ [ 902.859312] env[62460]: value = "task-1313805" [ 902.859312] env[62460]: _type = "Task" [ 902.859312] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.868255] env[62460]: DEBUG oslo_vmware.api [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313805, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.887053] env[62460]: DEBUG oslo_vmware.api [None req-8f649042-5fd7-4f94-80db-d6a078b9ce5a tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Task: {'id': task-1313803, 'name': PowerOffVM_Task, 'duration_secs': 0.193118} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.887360] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f649042-5fd7-4f94-80db-d6a078b9ce5a tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 902.887547] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-8f649042-5fd7-4f94-80db-d6a078b9ce5a tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 902.887794] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-69f8fabd-81c9-48da-b5f8-69d18663f99c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.919838] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Creating linked-clone VM from snapshot {{(pid=62460) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 902.920748] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-12ac6003-1dbf-40e0-9753-43115b3465a7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.929398] env[62460]: DEBUG oslo_vmware.api [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 902.929398] env[62460]: value = "task-1313807" [ 902.929398] env[62460]: _type = "Task" [ 902.929398] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.937783] env[62460]: DEBUG oslo_vmware.api [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313807, 'name': CloneVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.950190] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-8f649042-5fd7-4f94-80db-d6a078b9ce5a tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 902.950423] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-8f649042-5fd7-4f94-80db-d6a078b9ce5a tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Deleting contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 902.950625] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f649042-5fd7-4f94-80db-d6a078b9ce5a tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Deleting the datastore file [datastore1] 8bcaa672-97c8-466a-812c-5723e08d0c52 {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 902.950886] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f001f796-545a-42b1-9d20-70d34eaf7b67 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.960951] env[62460]: DEBUG oslo_vmware.api [None req-8f649042-5fd7-4f94-80db-d6a078b9ce5a tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Waiting for the task: (returnval){ [ 902.960951] env[62460]: value = "task-1313808" [ 902.960951] env[62460]: _type = "Task" [ 902.960951] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.969334] env[62460]: DEBUG oslo_vmware.api [None req-8f649042-5fd7-4f94-80db-d6a078b9ce5a tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Task: {'id': task-1313808, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.064197] env[62460]: DEBUG nova.network.neutron [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 903.069420] env[62460]: DEBUG nova.scheduler.client.report [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Updated inventory for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with generation 99 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 903.069871] env[62460]: DEBUG nova.compute.provider_tree [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Updating resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 generation from 99 to 100 during operation: update_inventory {{(pid=62460) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 903.070135] env[62460]: DEBUG nova.compute.provider_tree [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 903.209858] env[62460]: DEBUG oslo_vmware.api [None req-07205f12-9c2f-4768-9b9d-0f27bbfe3d9c tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Task: {'id': task-1313804, 'name': ReconfigVM_Task, 'duration_secs': 0.269972} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.210433] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-07205f12-9c2f-4768-9b9d-0f27bbfe3d9c tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Reconfigured VM instance instance-0000004b to detach disk 2000 {{(pid=62460) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 903.218990] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-22eeff13-1086-4a61-9b14-b77e75b04b00 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.237999] env[62460]: DEBUG nova.network.neutron [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Updating instance_info_cache with network_info: [{"id": "dbd267c2-b2d8-46ac-8d82-f645dd8560d5", "address": "fa:16:3e:a1:81:48", "network": {"id": "de789887-9cf9-4614-a7e1-dab430aaabaf", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-656400207-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0da056d93bdf40c39d6e82e457727ff6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5116f690-f825-4fee-8a47-42b073e716c5", "external-id": "nsx-vlan-transportzone-692", "segmentation_id": 692, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdbd267c2-b2", "ovs_interfaceid": "dbd267c2-b2d8-46ac-8d82-f645dd8560d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.246106] env[62460]: DEBUG oslo_vmware.api [None req-07205f12-9c2f-4768-9b9d-0f27bbfe3d9c tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Waiting for the task: (returnval){ [ 903.246106] env[62460]: value = "task-1313809" [ 903.246106] env[62460]: _type = "Task" [ 903.246106] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.259555] env[62460]: DEBUG oslo_vmware.api [None req-07205f12-9c2f-4768-9b9d-0f27bbfe3d9c tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Task: {'id': task-1313809, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.369600] env[62460]: DEBUG oslo_vmware.api [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313805, 'name': PowerOnVM_Task, 'duration_secs': 0.486438} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.369841] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 903.370074] env[62460]: INFO nova.compute.manager [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] Took 7.86 seconds to spawn the instance on the hypervisor. [ 903.370271] env[62460]: DEBUG nova.compute.manager [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 903.371098] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e134f67-e9f5-4359-869f-d57b5a9027da {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.438356] env[62460]: DEBUG oslo_vmware.api [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313807, 'name': CloneVM_Task} progress is 94%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.470532] env[62460]: DEBUG oslo_vmware.api [None req-8f649042-5fd7-4f94-80db-d6a078b9ce5a tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Task: {'id': task-1313808, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.168841} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.470799] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f649042-5fd7-4f94-80db-d6a078b9ce5a tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 903.470989] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-8f649042-5fd7-4f94-80db-d6a078b9ce5a tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Deleted contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 903.471202] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-8f649042-5fd7-4f94-80db-d6a078b9ce5a tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 903.471390] env[62460]: INFO nova.compute.manager [None req-8f649042-5fd7-4f94-80db-d6a078b9ce5a tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Took 1.11 seconds to destroy the instance on the hypervisor. [ 903.471647] env[62460]: DEBUG oslo.service.loopingcall [None req-8f649042-5fd7-4f94-80db-d6a078b9ce5a tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 903.471846] env[62460]: DEBUG nova.compute.manager [-] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 903.471945] env[62460]: DEBUG nova.network.neutron [-] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 903.575049] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.095s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.575747] env[62460]: DEBUG nova.compute.manager [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 903.578766] env[62460]: DEBUG oslo_concurrency.lockutils [None req-df5627bc-3771-4024-8583-4cb37ab37ea8 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.563s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.579140] env[62460]: DEBUG nova.objects.instance [None req-df5627bc-3771-4024-8583-4cb37ab37ea8 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lazy-loading 'resources' on Instance uuid fd03509c-5471-48bb-8150-d29531dca848 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 903.741051] env[62460]: DEBUG oslo_concurrency.lockutils [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Releasing lock "refresh_cache-09155629-51c8-4043-b1ee-6d5036552a67" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.741404] env[62460]: DEBUG nova.compute.manager [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Instance network_info: |[{"id": "dbd267c2-b2d8-46ac-8d82-f645dd8560d5", "address": "fa:16:3e:a1:81:48", "network": {"id": "de789887-9cf9-4614-a7e1-dab430aaabaf", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-656400207-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0da056d93bdf40c39d6e82e457727ff6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5116f690-f825-4fee-8a47-42b073e716c5", "external-id": "nsx-vlan-transportzone-692", "segmentation_id": 692, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdbd267c2-b2", "ovs_interfaceid": "dbd267c2-b2d8-46ac-8d82-f645dd8560d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 903.742165] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a1:81:48', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5116f690-f825-4fee-8a47-42b073e716c5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dbd267c2-b2d8-46ac-8d82-f645dd8560d5', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 903.751137] env[62460]: DEBUG oslo.service.loopingcall [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 903.751904] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 903.757414] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5e0e01ea-63e0-45c0-8a33-0f647f87b946 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.786199] env[62460]: DEBUG oslo_vmware.api [None req-07205f12-9c2f-4768-9b9d-0f27bbfe3d9c tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Task: {'id': task-1313809, 'name': ReconfigVM_Task, 'duration_secs': 0.177813} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.787600] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-07205f12-9c2f-4768-9b9d-0f27bbfe3d9c tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281168', 'volume_id': '2cf093c7-9278-4f2d-a385-8775acc7dae8', 'name': 'volume-2cf093c7-9278-4f2d-a385-8775acc7dae8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f1921907-f69c-49a3-896c-a4b12b526ffa', 'attached_at': '', 'detached_at': '', 'volume_id': '2cf093c7-9278-4f2d-a385-8775acc7dae8', 'serial': '2cf093c7-9278-4f2d-a385-8775acc7dae8'} {{(pid=62460) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 903.787929] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-07205f12-9c2f-4768-9b9d-0f27bbfe3d9c tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 903.788239] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 903.788239] env[62460]: value = "task-1313810" [ 903.788239] env[62460]: _type = "Task" [ 903.788239] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.789053] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91009603-aa78-41cd-bc5c-2af4da6b2cb4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.799781] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313810, 'name': CreateVM_Task} progress is 6%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.801937] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-07205f12-9c2f-4768-9b9d-0f27bbfe3d9c tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 903.802216] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3524b03d-2f6b-4265-9136-2415102b6ae6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.871190] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-07205f12-9c2f-4768-9b9d-0f27bbfe3d9c tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 903.871755] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-07205f12-9c2f-4768-9b9d-0f27bbfe3d9c tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Deleting contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 903.871934] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-07205f12-9c2f-4768-9b9d-0f27bbfe3d9c tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Deleting the datastore file [datastore2] f1921907-f69c-49a3-896c-a4b12b526ffa {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 903.872364] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3a66e802-1adf-4163-9393-d9e827fa926c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.879806] env[62460]: DEBUG oslo_vmware.api [None req-07205f12-9c2f-4768-9b9d-0f27bbfe3d9c tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Waiting for the task: (returnval){ [ 903.879806] env[62460]: value = "task-1313812" [ 903.879806] env[62460]: _type = "Task" [ 903.879806] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.889833] env[62460]: INFO nova.compute.manager [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] Took 30.11 seconds to build instance. [ 903.894328] env[62460]: DEBUG oslo_vmware.api [None req-07205f12-9c2f-4768-9b9d-0f27bbfe3d9c tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Task: {'id': task-1313812, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.940148] env[62460]: DEBUG oslo_vmware.api [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313807, 'name': CloneVM_Task} progress is 95%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.980402] env[62460]: DEBUG nova.compute.manager [req-054dea44-5f45-49a7-9e4a-38bd45b6001e req-3ef968fb-4175-4c26-90ee-0a03fe54ab66 service nova] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Received event network-changed-dbd267c2-b2d8-46ac-8d82-f645dd8560d5 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 903.980617] env[62460]: DEBUG nova.compute.manager [req-054dea44-5f45-49a7-9e4a-38bd45b6001e req-3ef968fb-4175-4c26-90ee-0a03fe54ab66 service nova] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Refreshing instance network info cache due to event network-changed-dbd267c2-b2d8-46ac-8d82-f645dd8560d5. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 903.980792] env[62460]: DEBUG oslo_concurrency.lockutils [req-054dea44-5f45-49a7-9e4a-38bd45b6001e req-3ef968fb-4175-4c26-90ee-0a03fe54ab66 service nova] Acquiring lock "refresh_cache-09155629-51c8-4043-b1ee-6d5036552a67" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 903.980947] env[62460]: DEBUG oslo_concurrency.lockutils [req-054dea44-5f45-49a7-9e4a-38bd45b6001e req-3ef968fb-4175-4c26-90ee-0a03fe54ab66 service nova] Acquired lock "refresh_cache-09155629-51c8-4043-b1ee-6d5036552a67" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.981133] env[62460]: DEBUG nova.network.neutron [req-054dea44-5f45-49a7-9e4a-38bd45b6001e req-3ef968fb-4175-4c26-90ee-0a03fe54ab66 service nova] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Refreshing network info cache for port dbd267c2-b2d8-46ac-8d82-f645dd8560d5 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 904.082592] env[62460]: DEBUG nova.compute.utils [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 904.084788] env[62460]: DEBUG nova.compute.manager [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 904.084930] env[62460]: DEBUG nova.network.neutron [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 904.149028] env[62460]: DEBUG nova.policy [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e531b7c3e7544e53b126d1336ac44c99', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bfa5a65b0f614e769de5b3aa77bf869e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 904.228183] env[62460]: DEBUG nova.network.neutron [-] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 904.280064] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bc2b005-e0db-4074-b793-8a9008ba3df1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.287923] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccd88649-02c8-465f-acf8-490f2f325a61 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.322652] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ef8a8cf-6e2e-457b-99f7-7f950323c496 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.328614] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313810, 'name': CreateVM_Task, 'duration_secs': 0.408076} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.329513] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 904.330597] env[62460]: DEBUG oslo_concurrency.lockutils [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.330597] env[62460]: DEBUG oslo_concurrency.lockutils [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.330597] env[62460]: DEBUG oslo_concurrency.lockutils [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 904.332873] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-44aae57d-e1bc-4b6c-b84d-2274ea45880c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.335500] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0568319d-d050-4239-898b-0017cbb51e0b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.344282] env[62460]: DEBUG oslo_vmware.api [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 904.344282] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]527f3271-6707-e38b-c0ff-070ab4c9e878" [ 904.344282] env[62460]: _type = "Task" [ 904.344282] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.351402] env[62460]: DEBUG nova.compute.provider_tree [None req-df5627bc-3771-4024-8583-4cb37ab37ea8 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 904.360695] env[62460]: DEBUG oslo_vmware.api [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]527f3271-6707-e38b-c0ff-070ab4c9e878, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.390270] env[62460]: DEBUG oslo_vmware.api [None req-07205f12-9c2f-4768-9b9d-0f27bbfe3d9c tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Task: {'id': task-1313812, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.098982} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.390672] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-07205f12-9c2f-4768-9b9d-0f27bbfe3d9c tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 904.390809] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-07205f12-9c2f-4768-9b9d-0f27bbfe3d9c tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Deleted contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 904.391653] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-07205f12-9c2f-4768-9b9d-0f27bbfe3d9c tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 904.391653] env[62460]: INFO nova.compute.manager [None req-07205f12-9c2f-4768-9b9d-0f27bbfe3d9c tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Took 2.29 seconds to destroy the instance on the hypervisor. [ 904.391653] env[62460]: DEBUG oslo.service.loopingcall [None req-07205f12-9c2f-4768-9b9d-0f27bbfe3d9c tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 904.391889] env[62460]: DEBUG nova.compute.manager [-] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 904.391990] env[62460]: DEBUG nova.network.neutron [-] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 904.395587] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e4949104-2bb9-4d16-bb15-9c64680af7c0 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "e90e8c3a-d244-4d04-997e-c15ab1cb2009" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.670s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.441715] env[62460]: DEBUG oslo_vmware.api [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313807, 'name': CloneVM_Task, 'duration_secs': 1.296892} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.443846] env[62460]: INFO nova.virt.vmwareapi.vmops [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Created linked-clone VM from snapshot [ 904.444648] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8a63cb1-4171-4fdb-b071-f0e67c2cd414 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.452230] env[62460]: DEBUG nova.virt.vmwareapi.images [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Uploading image 99119861-894e-4add-bc8a-5414d6c25b72 {{(pid=62460) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 904.475121] env[62460]: DEBUG oslo_vmware.rw_handles [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 904.475121] env[62460]: value = "vm-281233" [ 904.475121] env[62460]: _type = "VirtualMachine" [ 904.475121] env[62460]: }. {{(pid=62460) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 904.475486] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-492df7ab-7b05-4619-99a4-9a894fc5bc76 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.485036] env[62460]: DEBUG oslo_vmware.rw_handles [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lease: (returnval){ [ 904.485036] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52cfc9d2-3c2c-d683-09c2-cb592a8cf991" [ 904.485036] env[62460]: _type = "HttpNfcLease" [ 904.485036] env[62460]: } obtained for exporting VM: (result){ [ 904.485036] env[62460]: value = "vm-281233" [ 904.485036] env[62460]: _type = "VirtualMachine" [ 904.485036] env[62460]: }. {{(pid=62460) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 904.485354] env[62460]: DEBUG oslo_vmware.api [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the lease: (returnval){ [ 904.485354] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52cfc9d2-3c2c-d683-09c2-cb592a8cf991" [ 904.485354] env[62460]: _type = "HttpNfcLease" [ 904.485354] env[62460]: } to be ready. {{(pid=62460) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 904.494016] env[62460]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 904.494016] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52cfc9d2-3c2c-d683-09c2-cb592a8cf991" [ 904.494016] env[62460]: _type = "HttpNfcLease" [ 904.494016] env[62460]: } is ready. {{(pid=62460) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 904.494312] env[62460]: DEBUG oslo_vmware.rw_handles [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 904.494312] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52cfc9d2-3c2c-d683-09c2-cb592a8cf991" [ 904.494312] env[62460]: _type = "HttpNfcLease" [ 904.494312] env[62460]: }. {{(pid=62460) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 904.495053] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9961504f-196f-478e-b769-0690c53c8424 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.502705] env[62460]: DEBUG oslo_vmware.rw_handles [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52d24d14-774b-5856-cf94-c9e8f037af68/disk-0.vmdk from lease info. {{(pid=62460) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 904.502905] env[62460]: DEBUG oslo_vmware.rw_handles [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52d24d14-774b-5856-cf94-c9e8f037af68/disk-0.vmdk for reading. {{(pid=62460) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 904.573377] env[62460]: DEBUG nova.network.neutron [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Successfully created port: 3eaf04a4-feda-4a00-908e-d121fc775ecb {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 904.590805] env[62460]: DEBUG nova.compute.manager [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 904.610586] env[62460]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-b7bc36c2-262b-41f0-88e9-e14fd123bd59 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.731515] env[62460]: INFO nova.compute.manager [-] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Took 1.26 seconds to deallocate network for instance. [ 904.815204] env[62460]: DEBUG nova.network.neutron [req-054dea44-5f45-49a7-9e4a-38bd45b6001e req-3ef968fb-4175-4c26-90ee-0a03fe54ab66 service nova] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Updated VIF entry in instance network info cache for port dbd267c2-b2d8-46ac-8d82-f645dd8560d5. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 904.815774] env[62460]: DEBUG nova.network.neutron [req-054dea44-5f45-49a7-9e4a-38bd45b6001e req-3ef968fb-4175-4c26-90ee-0a03fe54ab66 service nova] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Updating instance_info_cache with network_info: [{"id": "dbd267c2-b2d8-46ac-8d82-f645dd8560d5", "address": "fa:16:3e:a1:81:48", "network": {"id": "de789887-9cf9-4614-a7e1-dab430aaabaf", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-656400207-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0da056d93bdf40c39d6e82e457727ff6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5116f690-f825-4fee-8a47-42b073e716c5", "external-id": "nsx-vlan-transportzone-692", "segmentation_id": 692, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdbd267c2-b2", "ovs_interfaceid": "dbd267c2-b2d8-46ac-8d82-f645dd8560d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 904.867290] env[62460]: DEBUG oslo_vmware.api [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]527f3271-6707-e38b-c0ff-070ab4c9e878, 'name': SearchDatastore_Task, 'duration_secs': 0.016828} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.867290] env[62460]: DEBUG oslo_concurrency.lockutils [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.867290] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 904.867290] env[62460]: DEBUG oslo_concurrency.lockutils [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.867290] env[62460]: DEBUG oslo_concurrency.lockutils [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.867290] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 904.867290] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2d21cfff-b01e-4f31-89aa-8a5448c33143 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.875476] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 904.875476] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 904.875790] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-df6db6c6-a364-45fe-a164-e5fd554103e1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.879705] env[62460]: ERROR nova.scheduler.client.report [None req-df5627bc-3771-4024-8583-4cb37ab37ea8 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [req-49b003c4-3a51-4d41-b723-f954a8ffa26e] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 76e8f00d-65cc-4766-bb9d-5006fb51c728. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-49b003c4-3a51-4d41-b723-f954a8ffa26e"}]} [ 904.883759] env[62460]: DEBUG oslo_vmware.api [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 904.883759] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]5272a818-33d8-7490-ec17-5be21943a136" [ 904.883759] env[62460]: _type = "Task" [ 904.883759] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.892136] env[62460]: DEBUG oslo_vmware.api [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5272a818-33d8-7490-ec17-5be21943a136, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.905155] env[62460]: DEBUG nova.scheduler.client.report [None req-df5627bc-3771-4024-8583-4cb37ab37ea8 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Refreshing inventories for resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 904.922588] env[62460]: DEBUG nova.scheduler.client.report [None req-df5627bc-3771-4024-8583-4cb37ab37ea8 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Updating ProviderTree inventory for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 904.922827] env[62460]: DEBUG nova.compute.provider_tree [None req-df5627bc-3771-4024-8583-4cb37ab37ea8 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 904.947254] env[62460]: DEBUG nova.scheduler.client.report [None req-df5627bc-3771-4024-8583-4cb37ab37ea8 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Refreshing aggregate associations for resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728, aggregates: None {{(pid=62460) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 904.971952] env[62460]: DEBUG nova.scheduler.client.report [None req-df5627bc-3771-4024-8583-4cb37ab37ea8 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Refreshing trait associations for resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62460) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 905.200061] env[62460]: DEBUG oslo_concurrency.lockutils [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "61073a4e-c5e5-48f5-a43a-ce71bd18e48a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.200861] env[62460]: DEBUG oslo_concurrency.lockutils [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "61073a4e-c5e5-48f5-a43a-ce71bd18e48a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.240934] env[62460]: DEBUG oslo_concurrency.lockutils [None req-8f649042-5fd7-4f94-80db-d6a078b9ce5a tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.260586] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c17409b-592d-48dc-b4c1-2111460b77ba {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.269494] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-558aa3b7-0f7d-4ae2-b8e8-0b609c1b6e71 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.301257] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27306bb0-665d-484a-b167-9b8bb6f000c7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.309662] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50527312-2957-4cff-8704-d86dd097a971 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.324358] env[62460]: DEBUG oslo_concurrency.lockutils [req-054dea44-5f45-49a7-9e4a-38bd45b6001e req-3ef968fb-4175-4c26-90ee-0a03fe54ab66 service nova] Releasing lock "refresh_cache-09155629-51c8-4043-b1ee-6d5036552a67" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.324795] env[62460]: DEBUG nova.compute.manager [req-054dea44-5f45-49a7-9e4a-38bd45b6001e req-3ef968fb-4175-4c26-90ee-0a03fe54ab66 service nova] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Received event network-vif-deleted-797761fb-cacf-47d3-8617-691a88101883 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 905.325142] env[62460]: INFO nova.compute.manager [req-054dea44-5f45-49a7-9e4a-38bd45b6001e req-3ef968fb-4175-4c26-90ee-0a03fe54ab66 service nova] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Neutron deleted interface 797761fb-cacf-47d3-8617-691a88101883; detaching it from the instance and deleting it from the info cache [ 905.325408] env[62460]: DEBUG nova.network.neutron [req-054dea44-5f45-49a7-9e4a-38bd45b6001e req-3ef968fb-4175-4c26-90ee-0a03fe54ab66 service nova] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.327427] env[62460]: DEBUG nova.compute.provider_tree [None req-df5627bc-3771-4024-8583-4cb37ab37ea8 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 905.394053] env[62460]: DEBUG oslo_vmware.api [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5272a818-33d8-7490-ec17-5be21943a136, 'name': SearchDatastore_Task, 'duration_secs': 0.01187} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.394979] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2bcfd6f2-d9a7-4574-8f5f-b3a43e317489 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.402561] env[62460]: DEBUG oslo_vmware.api [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 905.402561] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]524c09fc-a14f-ec59-8051-137c06f44c0b" [ 905.402561] env[62460]: _type = "Task" [ 905.402561] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.412284] env[62460]: DEBUG oslo_vmware.api [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]524c09fc-a14f-ec59-8051-137c06f44c0b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.604285] env[62460]: DEBUG nova.compute.manager [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 905.625878] env[62460]: DEBUG nova.virt.hardware [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 905.626285] env[62460]: DEBUG nova.virt.hardware [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 905.626601] env[62460]: DEBUG nova.virt.hardware [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 905.626849] env[62460]: DEBUG nova.virt.hardware [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 905.627105] env[62460]: DEBUG nova.virt.hardware [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 905.627373] env[62460]: DEBUG nova.virt.hardware [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 905.627612] env[62460]: DEBUG nova.virt.hardware [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 905.627909] env[62460]: DEBUG nova.virt.hardware [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 905.628104] env[62460]: DEBUG nova.virt.hardware [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 905.628289] env[62460]: DEBUG nova.virt.hardware [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 905.628474] env[62460]: DEBUG nova.virt.hardware [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 905.629352] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a984115-64b6-499c-b0f7-2183bde74fea {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.645223] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9acba2a0-15bf-4e88-a0ca-95add2a8e387 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.659558] env[62460]: DEBUG nova.network.neutron [-] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.704455] env[62460]: DEBUG nova.compute.manager [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 905.833717] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d46b51e5-59a6-4321-9d29-5b84c70fe854 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.843405] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83e484ac-f927-4758-a0aa-55908dcea8d0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.859954] env[62460]: DEBUG nova.scheduler.client.report [None req-df5627bc-3771-4024-8583-4cb37ab37ea8 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Updated inventory for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with generation 101 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 905.860334] env[62460]: DEBUG nova.compute.provider_tree [None req-df5627bc-3771-4024-8583-4cb37ab37ea8 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Updating resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 generation from 101 to 102 during operation: update_inventory {{(pid=62460) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 905.860622] env[62460]: DEBUG nova.compute.provider_tree [None req-df5627bc-3771-4024-8583-4cb37ab37ea8 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 905.871481] env[62460]: DEBUG nova.compute.manager [req-054dea44-5f45-49a7-9e4a-38bd45b6001e req-3ef968fb-4175-4c26-90ee-0a03fe54ab66 service nova] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Detach interface failed, port_id=797761fb-cacf-47d3-8617-691a88101883, reason: Instance 8bcaa672-97c8-466a-812c-5723e08d0c52 could not be found. {{(pid=62460) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 905.913923] env[62460]: DEBUG oslo_vmware.api [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]524c09fc-a14f-ec59-8051-137c06f44c0b, 'name': SearchDatastore_Task, 'duration_secs': 0.017997} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.914334] env[62460]: DEBUG oslo_concurrency.lockutils [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.914659] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] 09155629-51c8-4043-b1ee-6d5036552a67/09155629-51c8-4043-b1ee-6d5036552a67.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 905.914990] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-79bf24b0-b8a4-4c74-964c-722b43017d55 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.922052] env[62460]: DEBUG oslo_vmware.api [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 905.922052] env[62460]: value = "task-1313814" [ 905.922052] env[62460]: _type = "Task" [ 905.922052] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.930492] env[62460]: DEBUG oslo_vmware.api [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313814, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.005695] env[62460]: DEBUG nova.compute.manager [req-c4e8753b-e59a-48d2-878e-2f8f7d88d5a2 req-a7f52058-4656-4653-93cd-8f88e916a013 service nova] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Received event network-vif-deleted-dbacff36-64b1-4fc7-b0a0-dc78fcb72ac2 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 906.162719] env[62460]: INFO nova.compute.manager [-] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Took 1.77 seconds to deallocate network for instance. [ 906.240461] env[62460]: DEBUG oslo_concurrency.lockutils [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.373660] env[62460]: DEBUG oslo_concurrency.lockutils [None req-df5627bc-3771-4024-8583-4cb37ab37ea8 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.795s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.376422] env[62460]: DEBUG oslo_concurrency.lockutils [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.250s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.378366] env[62460]: INFO nova.compute.claims [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 906.394564] env[62460]: INFO nova.scheduler.client.report [None req-df5627bc-3771-4024-8583-4cb37ab37ea8 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Deleted allocations for instance fd03509c-5471-48bb-8150-d29531dca848 [ 906.436683] env[62460]: DEBUG oslo_vmware.api [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313814, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.465725] env[62460]: DEBUG nova.network.neutron [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Successfully updated port: 3eaf04a4-feda-4a00-908e-d121fc775ecb {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 906.720029] env[62460]: INFO nova.compute.manager [None req-07205f12-9c2f-4768-9b9d-0f27bbfe3d9c tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Took 0.56 seconds to detach 1 volumes for instance. [ 906.722219] env[62460]: DEBUG nova.compute.manager [None req-07205f12-9c2f-4768-9b9d-0f27bbfe3d9c tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Deleting volume: 2cf093c7-9278-4f2d-a385-8775acc7dae8 {{(pid=62460) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 906.902616] env[62460]: DEBUG oslo_concurrency.lockutils [None req-df5627bc-3771-4024-8583-4cb37ab37ea8 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "fd03509c-5471-48bb-8150-d29531dca848" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.911s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.932671] env[62460]: DEBUG oslo_vmware.api [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313814, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.572276} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.932671] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] 09155629-51c8-4043-b1ee-6d5036552a67/09155629-51c8-4043-b1ee-6d5036552a67.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 906.932671] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 906.932956] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5d311b04-d3c8-4c27-8385-7c96aa0da343 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.939333] env[62460]: DEBUG oslo_vmware.api [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 906.939333] env[62460]: value = "task-1313816" [ 906.939333] env[62460]: _type = "Task" [ 906.939333] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.948447] env[62460]: DEBUG oslo_vmware.api [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313816, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.968937] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "refresh_cache-6b084ed4-e8f0-4063-bf2a-252740640753" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.969089] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquired lock "refresh_cache-6b084ed4-e8f0-4063-bf2a-252740640753" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.969251] env[62460]: DEBUG nova.network.neutron [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 907.263740] env[62460]: DEBUG oslo_concurrency.lockutils [None req-07205f12-9c2f-4768-9b9d-0f27bbfe3d9c tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.450589] env[62460]: DEBUG oslo_vmware.api [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313816, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065843} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.453392] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 907.457182] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2a730a4-73c6-4f8a-94ba-edcc61f66831 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.480790] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Reconfiguring VM instance instance-00000054 to attach disk [datastore2] 09155629-51c8-4043-b1ee-6d5036552a67/09155629-51c8-4043-b1ee-6d5036552a67.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 907.486480] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3efbade4-aa91-4166-9bb9-985a4b387e8f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.513130] env[62460]: DEBUG oslo_vmware.api [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 907.513130] env[62460]: value = "task-1313817" [ 907.513130] env[62460]: _type = "Task" [ 907.513130] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.526206] env[62460]: DEBUG oslo_vmware.api [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313817, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.532776] env[62460]: DEBUG nova.network.neutron [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 907.613874] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ccd6a6a-122b-41b1-acb5-53fb487785e9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.621904] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec00b6c9-3453-4e20-a72f-1e7b9dceeda2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.657763] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ffeb425-f79c-4364-9f88-ebdbed3824bc {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.665678] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03977a5e-4e1e-481f-a9b1-8966939844fd {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.679504] env[62460]: DEBUG nova.compute.provider_tree [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 907.697072] env[62460]: DEBUG nova.network.neutron [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Updating instance_info_cache with network_info: [{"id": "3eaf04a4-feda-4a00-908e-d121fc775ecb", "address": "fa:16:3e:98:27:28", "network": {"id": "3c49720c-94c9-4e2f-bbb7-4266cc97faee", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1650281790-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfa5a65b0f614e769de5b3aa77bf869e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3eaf04a4-fe", "ovs_interfaceid": "3eaf04a4-feda-4a00-908e-d121fc775ecb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 908.027705] env[62460]: DEBUG oslo_vmware.api [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313817, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.032596] env[62460]: DEBUG nova.compute.manager [req-dfc2cc69-4821-4a6c-8a95-f27fcf7dca55 req-caa4cf72-86e0-4aa7-ac65-0587e26c77e4 service nova] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Received event network-vif-plugged-3eaf04a4-feda-4a00-908e-d121fc775ecb {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 908.032858] env[62460]: DEBUG oslo_concurrency.lockutils [req-dfc2cc69-4821-4a6c-8a95-f27fcf7dca55 req-caa4cf72-86e0-4aa7-ac65-0587e26c77e4 service nova] Acquiring lock "6b084ed4-e8f0-4063-bf2a-252740640753-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.033228] env[62460]: DEBUG oslo_concurrency.lockutils [req-dfc2cc69-4821-4a6c-8a95-f27fcf7dca55 req-caa4cf72-86e0-4aa7-ac65-0587e26c77e4 service nova] Lock "6b084ed4-e8f0-4063-bf2a-252740640753-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.033476] env[62460]: DEBUG oslo_concurrency.lockutils [req-dfc2cc69-4821-4a6c-8a95-f27fcf7dca55 req-caa4cf72-86e0-4aa7-ac65-0587e26c77e4 service nova] Lock "6b084ed4-e8f0-4063-bf2a-252740640753-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.033675] env[62460]: DEBUG nova.compute.manager [req-dfc2cc69-4821-4a6c-8a95-f27fcf7dca55 req-caa4cf72-86e0-4aa7-ac65-0587e26c77e4 service nova] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] No waiting events found dispatching network-vif-plugged-3eaf04a4-feda-4a00-908e-d121fc775ecb {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 908.033944] env[62460]: WARNING nova.compute.manager [req-dfc2cc69-4821-4a6c-8a95-f27fcf7dca55 req-caa4cf72-86e0-4aa7-ac65-0587e26c77e4 service nova] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Received unexpected event network-vif-plugged-3eaf04a4-feda-4a00-908e-d121fc775ecb for instance with vm_state building and task_state spawning. [ 908.034184] env[62460]: DEBUG nova.compute.manager [req-dfc2cc69-4821-4a6c-8a95-f27fcf7dca55 req-caa4cf72-86e0-4aa7-ac65-0587e26c77e4 service nova] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Received event network-changed-3eaf04a4-feda-4a00-908e-d121fc775ecb {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 908.034396] env[62460]: DEBUG nova.compute.manager [req-dfc2cc69-4821-4a6c-8a95-f27fcf7dca55 req-caa4cf72-86e0-4aa7-ac65-0587e26c77e4 service nova] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Refreshing instance network info cache due to event network-changed-3eaf04a4-feda-4a00-908e-d121fc775ecb. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 908.034585] env[62460]: DEBUG oslo_concurrency.lockutils [req-dfc2cc69-4821-4a6c-8a95-f27fcf7dca55 req-caa4cf72-86e0-4aa7-ac65-0587e26c77e4 service nova] Acquiring lock "refresh_cache-6b084ed4-e8f0-4063-bf2a-252740640753" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 908.182924] env[62460]: DEBUG nova.scheduler.client.report [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 908.199289] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Releasing lock "refresh_cache-6b084ed4-e8f0-4063-bf2a-252740640753" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.199595] env[62460]: DEBUG nova.compute.manager [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Instance network_info: |[{"id": "3eaf04a4-feda-4a00-908e-d121fc775ecb", "address": "fa:16:3e:98:27:28", "network": {"id": "3c49720c-94c9-4e2f-bbb7-4266cc97faee", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1650281790-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfa5a65b0f614e769de5b3aa77bf869e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3eaf04a4-fe", "ovs_interfaceid": "3eaf04a4-feda-4a00-908e-d121fc775ecb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 908.200210] env[62460]: DEBUG oslo_concurrency.lockutils [req-dfc2cc69-4821-4a6c-8a95-f27fcf7dca55 req-caa4cf72-86e0-4aa7-ac65-0587e26c77e4 service nova] Acquired lock "refresh_cache-6b084ed4-e8f0-4063-bf2a-252740640753" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.200411] env[62460]: DEBUG nova.network.neutron [req-dfc2cc69-4821-4a6c-8a95-f27fcf7dca55 req-caa4cf72-86e0-4aa7-ac65-0587e26c77e4 service nova] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Refreshing network info cache for port 3eaf04a4-feda-4a00-908e-d121fc775ecb {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 908.201641] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:98:27:28', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '04ccbc7a-cf8d-4ea2-8411-291a1e27df7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3eaf04a4-feda-4a00-908e-d121fc775ecb', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 908.209696] env[62460]: DEBUG oslo.service.loopingcall [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 908.210748] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 908.211047] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-197fef8d-b526-4aa5-8d81-ba62e39dba65 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.231581] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 908.231581] env[62460]: value = "task-1313818" [ 908.231581] env[62460]: _type = "Task" [ 908.231581] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.240140] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313818, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.527733] env[62460]: DEBUG oslo_vmware.api [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313817, 'name': ReconfigVM_Task, 'duration_secs': 0.672918} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.528104] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Reconfigured VM instance instance-00000054 to attach disk [datastore2] 09155629-51c8-4043-b1ee-6d5036552a67/09155629-51c8-4043-b1ee-6d5036552a67.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 908.528728] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7dcb7534-8f79-4a69-bd46-97bb773bfb6c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.536975] env[62460]: DEBUG oslo_vmware.api [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 908.536975] env[62460]: value = "task-1313819" [ 908.536975] env[62460]: _type = "Task" [ 908.536975] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.547374] env[62460]: DEBUG oslo_vmware.api [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313819, 'name': Rename_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.691024] env[62460]: DEBUG oslo_concurrency.lockutils [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.312s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.691024] env[62460]: DEBUG nova.compute.manager [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 908.692225] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 11.465s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.744567] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313818, 'name': CreateVM_Task} progress is 99%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.939228] env[62460]: DEBUG nova.network.neutron [req-dfc2cc69-4821-4a6c-8a95-f27fcf7dca55 req-caa4cf72-86e0-4aa7-ac65-0587e26c77e4 service nova] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Updated VIF entry in instance network info cache for port 3eaf04a4-feda-4a00-908e-d121fc775ecb. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 908.939612] env[62460]: DEBUG nova.network.neutron [req-dfc2cc69-4821-4a6c-8a95-f27fcf7dca55 req-caa4cf72-86e0-4aa7-ac65-0587e26c77e4 service nova] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Updating instance_info_cache with network_info: [{"id": "3eaf04a4-feda-4a00-908e-d121fc775ecb", "address": "fa:16:3e:98:27:28", "network": {"id": "3c49720c-94c9-4e2f-bbb7-4266cc97faee", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1650281790-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfa5a65b0f614e769de5b3aa77bf869e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3eaf04a4-fe", "ovs_interfaceid": "3eaf04a4-feda-4a00-908e-d121fc775ecb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 909.048227] env[62460]: DEBUG oslo_vmware.api [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313819, 'name': Rename_Task} progress is 99%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.196348] env[62460]: DEBUG nova.compute.utils [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 909.203042] env[62460]: DEBUG nova.compute.manager [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 909.203156] env[62460]: DEBUG nova.network.neutron [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 909.243412] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313818, 'name': CreateVM_Task} progress is 99%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.255359] env[62460]: DEBUG nova.policy [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '01d9aa372c4a43bda7148cb46081d0d1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1845e17fb7e64cf594c4e81c67ad8243', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 909.444676] env[62460]: DEBUG oslo_concurrency.lockutils [req-dfc2cc69-4821-4a6c-8a95-f27fcf7dca55 req-caa4cf72-86e0-4aa7-ac65-0587e26c77e4 service nova] Releasing lock "refresh_cache-6b084ed4-e8f0-4063-bf2a-252740640753" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.544308] env[62460]: DEBUG nova.network.neutron [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Successfully created port: 7bb00d79-430f-44e2-9e92-57109c05d0bb {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 909.552867] env[62460]: DEBUG oslo_vmware.api [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313819, 'name': Rename_Task} progress is 99%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.704686] env[62460]: DEBUG nova.compute.manager [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 909.742856] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313818, 'name': CreateVM_Task, 'duration_secs': 1.428474} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.743925] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 5b6e8205-003e-49c0-a73d-be2e032a8272 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 909.744107] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 896f8ff4-e45f-4403-a727-03ee25e58609 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 909.744267] env[62460]: WARNING nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance f1921907-f69c-49a3-896c-a4b12b526ffa is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 909.744394] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 3e10af70-db52-4d4a-bb92-821a05dcbab6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 909.744549] env[62460]: WARNING nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 0da70720-197c-4483-93e1-019e0f040c4c is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 909.744684] env[62460]: WARNING nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance fbe558f1-ce5a-4429-8833-48bdabe91071 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 909.744936] env[62460]: WARNING nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 8bcaa672-97c8-466a-812c-5723e08d0c52 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 909.745095] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance e90e8c3a-d244-4d04-997e-c15ab1cb2009 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 909.745348] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 09155629-51c8-4043-b1ee-6d5036552a67 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 909.745348] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 6b084ed4-e8f0-4063-bf2a-252740640753 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 909.745473] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance bf28fc98-d9a9-4678-a107-ace2b6503353 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 909.750021] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 909.750021] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.750021] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.750021] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 909.750021] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c6cf6c2a-39c9-491b-8b2a-d9b5ec680cc7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.756040] env[62460]: DEBUG oslo_vmware.api [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 909.756040] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52cd2f18-cd42-0296-2230-5d43aac6c697" [ 909.756040] env[62460]: _type = "Task" [ 909.756040] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.765273] env[62460]: DEBUG oslo_vmware.api [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52cd2f18-cd42-0296-2230-5d43aac6c697, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.051127] env[62460]: DEBUG oslo_vmware.api [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313819, 'name': Rename_Task, 'duration_secs': 1.188055} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.051812] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 910.052289] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-61cf5ef5-41ed-4daa-bdf9-6627c4259cd3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.059149] env[62460]: DEBUG oslo_vmware.api [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 910.059149] env[62460]: value = "task-1313820" [ 910.059149] env[62460]: _type = "Task" [ 910.059149] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.067180] env[62460]: DEBUG oslo_vmware.api [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313820, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.248509] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance abd4501b-3ad8-4103-921d-5b80fc2f313f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 910.269977] env[62460]: DEBUG oslo_vmware.api [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52cd2f18-cd42-0296-2230-5d43aac6c697, 'name': SearchDatastore_Task, 'duration_secs': 0.015572} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.270385] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 910.270637] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 910.270988] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.271190] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.271520] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 910.271831] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ecf4747d-dec3-4086-97eb-32b831c69c61 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.283326] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 910.283527] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 910.284991] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c8ba0d39-5679-4d66-b2c7-b878dd6454c7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.290096] env[62460]: DEBUG oslo_vmware.api [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 910.290096] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52307034-832c-8f8b-5c59-c20513146233" [ 910.290096] env[62460]: _type = "Task" [ 910.290096] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.298445] env[62460]: DEBUG oslo_vmware.api [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52307034-832c-8f8b-5c59-c20513146233, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.571099] env[62460]: DEBUG oslo_vmware.api [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313820, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.716016] env[62460]: DEBUG nova.compute.manager [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 910.741389] env[62460]: DEBUG nova.virt.hardware [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 910.741762] env[62460]: DEBUG nova.virt.hardware [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 910.741974] env[62460]: DEBUG nova.virt.hardware [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 910.742234] env[62460]: DEBUG nova.virt.hardware [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 910.742405] env[62460]: DEBUG nova.virt.hardware [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 910.742566] env[62460]: DEBUG nova.virt.hardware [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 910.742785] env[62460]: DEBUG nova.virt.hardware [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 910.742947] env[62460]: DEBUG nova.virt.hardware [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 910.743134] env[62460]: DEBUG nova.virt.hardware [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 910.743307] env[62460]: DEBUG nova.virt.hardware [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 910.743490] env[62460]: DEBUG nova.virt.hardware [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 910.744441] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83b6556d-963f-4860-9a49-be997a54125e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.751811] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 61073a4e-c5e5-48f5-a43a-ce71bd18e48a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 910.752079] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Total usable vcpus: 48, total allocated vcpus: 7 {{(pid=62460) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 910.752268] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1856MB phys_disk=200GB used_disk=7GB total_vcpus=48 used_vcpus=7 pci_stats=[] {{(pid=62460) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 910.755674] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d123b20-9d66-455c-9865-4c2a788e5773 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.802984] env[62460]: DEBUG oslo_vmware.api [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52307034-832c-8f8b-5c59-c20513146233, 'name': SearchDatastore_Task, 'duration_secs': 0.013818} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.803804] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e92d5b02-9973-4fb6-9c7b-8f6d3d3227da {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.808982] env[62460]: DEBUG oslo_vmware.api [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 910.808982] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52f6f73b-2751-e2b8-ffea-e45da9af9b3e" [ 910.808982] env[62460]: _type = "Task" [ 910.808982] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.817881] env[62460]: DEBUG oslo_vmware.api [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52f6f73b-2751-e2b8-ffea-e45da9af9b3e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.927726] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-099d5a55-6fb1-46bf-b0b0-5dcbcbb51dcd {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.935512] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b60c4627-9445-43a0-b930-1875f0055933 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.965271] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdf02a9b-ed6e-4865-ab8e-570962530b12 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.972662] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dcd4c8f-11eb-4313-b514-585c0d66a6cf {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.989198] env[62460]: DEBUG nova.compute.provider_tree [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 911.074017] env[62460]: DEBUG oslo_vmware.api [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313820, 'name': PowerOnVM_Task, 'duration_secs': 0.617064} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.074321] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 911.074534] env[62460]: INFO nova.compute.manager [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Took 9.59 seconds to spawn the instance on the hypervisor. [ 911.074725] env[62460]: DEBUG nova.compute.manager [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 911.078732] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d220fdf9-b1d0-4c72-ab14-e82533bb75f5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.264122] env[62460]: DEBUG nova.compute.manager [req-ac883301-58d7-4ce6-9032-abaa151a6b0b req-9428e824-4f6f-416f-bd1c-f01258938767 service nova] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Received event network-vif-plugged-7bb00d79-430f-44e2-9e92-57109c05d0bb {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 911.264404] env[62460]: DEBUG oslo_concurrency.lockutils [req-ac883301-58d7-4ce6-9032-abaa151a6b0b req-9428e824-4f6f-416f-bd1c-f01258938767 service nova] Acquiring lock "bf28fc98-d9a9-4678-a107-ace2b6503353-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.264657] env[62460]: DEBUG oslo_concurrency.lockutils [req-ac883301-58d7-4ce6-9032-abaa151a6b0b req-9428e824-4f6f-416f-bd1c-f01258938767 service nova] Lock "bf28fc98-d9a9-4678-a107-ace2b6503353-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.264857] env[62460]: DEBUG oslo_concurrency.lockutils [req-ac883301-58d7-4ce6-9032-abaa151a6b0b req-9428e824-4f6f-416f-bd1c-f01258938767 service nova] Lock "bf28fc98-d9a9-4678-a107-ace2b6503353-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.265236] env[62460]: DEBUG nova.compute.manager [req-ac883301-58d7-4ce6-9032-abaa151a6b0b req-9428e824-4f6f-416f-bd1c-f01258938767 service nova] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] No waiting events found dispatching network-vif-plugged-7bb00d79-430f-44e2-9e92-57109c05d0bb {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 911.265491] env[62460]: WARNING nova.compute.manager [req-ac883301-58d7-4ce6-9032-abaa151a6b0b req-9428e824-4f6f-416f-bd1c-f01258938767 service nova] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Received unexpected event network-vif-plugged-7bb00d79-430f-44e2-9e92-57109c05d0bb for instance with vm_state building and task_state spawning. [ 911.320545] env[62460]: DEBUG oslo_vmware.api [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52f6f73b-2751-e2b8-ffea-e45da9af9b3e, 'name': SearchDatastore_Task, 'duration_secs': 0.02022} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.320545] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.320545] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] 6b084ed4-e8f0-4063-bf2a-252740640753/6b084ed4-e8f0-4063-bf2a-252740640753.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 911.320545] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fe39f425-6923-4758-949b-b4cadcb99388 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.327634] env[62460]: DEBUG oslo_vmware.api [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 911.327634] env[62460]: value = "task-1313821" [ 911.327634] env[62460]: _type = "Task" [ 911.327634] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.336692] env[62460]: DEBUG oslo_vmware.api [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313821, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.361737] env[62460]: DEBUG nova.network.neutron [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Successfully updated port: 7bb00d79-430f-44e2-9e92-57109c05d0bb {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 911.492961] env[62460]: DEBUG nova.scheduler.client.report [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 911.595237] env[62460]: INFO nova.compute.manager [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Took 25.83 seconds to build instance. [ 911.838932] env[62460]: DEBUG oslo_vmware.api [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313821, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.864006] env[62460]: DEBUG oslo_concurrency.lockutils [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Acquiring lock "refresh_cache-bf28fc98-d9a9-4678-a107-ace2b6503353" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.864323] env[62460]: DEBUG oslo_concurrency.lockutils [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Acquired lock "refresh_cache-bf28fc98-d9a9-4678-a107-ace2b6503353" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.864565] env[62460]: DEBUG nova.network.neutron [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 911.998440] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62460) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 911.999128] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.307s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.000123] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.855s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.001153] env[62460]: INFO nova.compute.claims [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 912.062291] env[62460]: DEBUG oslo_vmware.rw_handles [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52d24d14-774b-5856-cf94-c9e8f037af68/disk-0.vmdk. {{(pid=62460) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 912.063217] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fb2861e-b4ba-407b-8216-4fca3b37357f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.070477] env[62460]: DEBUG oslo_vmware.rw_handles [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52d24d14-774b-5856-cf94-c9e8f037af68/disk-0.vmdk is in state: ready. {{(pid=62460) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 912.070656] env[62460]: ERROR oslo_vmware.rw_handles [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52d24d14-774b-5856-cf94-c9e8f037af68/disk-0.vmdk due to incomplete transfer. [ 912.071159] env[62460]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-0f42f5f8-db5d-48c9-9415-f2e36f92415e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.079212] env[62460]: DEBUG oslo_vmware.rw_handles [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52d24d14-774b-5856-cf94-c9e8f037af68/disk-0.vmdk. {{(pid=62460) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 912.079414] env[62460]: DEBUG nova.virt.vmwareapi.images [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Uploaded image 99119861-894e-4add-bc8a-5414d6c25b72 to the Glance image server {{(pid=62460) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 912.081547] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Destroying the VM {{(pid=62460) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 912.081795] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-bf5ae80a-b358-4c4c-b03f-632e0252bb63 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.087531] env[62460]: DEBUG oslo_vmware.api [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 912.087531] env[62460]: value = "task-1313822" [ 912.087531] env[62460]: _type = "Task" [ 912.087531] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.094923] env[62460]: DEBUG oslo_vmware.api [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313822, 'name': Destroy_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.097406] env[62460]: DEBUG oslo_concurrency.lockutils [None req-423e4c1d-86a1-4abc-ad44-16048232f1fe tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "09155629-51c8-4043-b1ee-6d5036552a67" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.341s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.338188] env[62460]: DEBUG oslo_vmware.api [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313821, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.681764} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.338464] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] 6b084ed4-e8f0-4063-bf2a-252740640753/6b084ed4-e8f0-4063-bf2a-252740640753.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 912.338828] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 912.339036] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ed04fb5d-6fd1-442f-bede-f858e7e796f9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.345635] env[62460]: DEBUG oslo_vmware.api [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 912.345635] env[62460]: value = "task-1313823" [ 912.345635] env[62460]: _type = "Task" [ 912.345635] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.353479] env[62460]: DEBUG oslo_vmware.api [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313823, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.405370] env[62460]: DEBUG nova.network.neutron [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 912.597815] env[62460]: DEBUG oslo_vmware.api [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313822, 'name': Destroy_Task} progress is 33%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.626681] env[62460]: DEBUG nova.network.neutron [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Updating instance_info_cache with network_info: [{"id": "7bb00d79-430f-44e2-9e92-57109c05d0bb", "address": "fa:16:3e:94:48:0c", "network": {"id": "e2ad0587-12a2-40ce-b2d0-c682042335fb", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1549110976-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1845e17fb7e64cf594c4e81c67ad8243", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c405e9f-a6c8-4308-acac-071654efe18e", "external-id": "nsx-vlan-transportzone-851", "segmentation_id": 851, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7bb00d79-43", "ovs_interfaceid": "7bb00d79-430f-44e2-9e92-57109c05d0bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.663146] env[62460]: DEBUG oslo_concurrency.lockutils [None req-48989aff-581d-4da7-8344-7d4e8beb1358 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "09155629-51c8-4043-b1ee-6d5036552a67" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.663508] env[62460]: DEBUG oslo_concurrency.lockutils [None req-48989aff-581d-4da7-8344-7d4e8beb1358 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "09155629-51c8-4043-b1ee-6d5036552a67" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.663627] env[62460]: DEBUG nova.compute.manager [None req-48989aff-581d-4da7-8344-7d4e8beb1358 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 912.664579] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bf60c05-8152-4d9d-9581-0d166a88d5a5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.671920] env[62460]: DEBUG nova.compute.manager [None req-48989aff-581d-4da7-8344-7d4e8beb1358 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62460) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 912.672586] env[62460]: DEBUG nova.objects.instance [None req-48989aff-581d-4da7-8344-7d4e8beb1358 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lazy-loading 'flavor' on Instance uuid 09155629-51c8-4043-b1ee-6d5036552a67 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 912.856477] env[62460]: DEBUG oslo_vmware.api [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313823, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.416593} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.856767] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 912.857641] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fb77cb3-e45e-4a99-91c9-ebf47e11311c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.879933] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Reconfiguring VM instance instance-00000055 to attach disk [datastore2] 6b084ed4-e8f0-4063-bf2a-252740640753/6b084ed4-e8f0-4063-bf2a-252740640753.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 912.880240] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a5c891ad-2abc-4be1-bb2d-38719b648694 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.899526] env[62460]: DEBUG oslo_vmware.api [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 912.899526] env[62460]: value = "task-1313824" [ 912.899526] env[62460]: _type = "Task" [ 912.899526] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.907753] env[62460]: DEBUG oslo_vmware.api [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313824, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.101082] env[62460]: DEBUG oslo_vmware.api [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313822, 'name': Destroy_Task} progress is 100%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.131744] env[62460]: DEBUG oslo_concurrency.lockutils [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Releasing lock "refresh_cache-bf28fc98-d9a9-4678-a107-ace2b6503353" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.131744] env[62460]: DEBUG nova.compute.manager [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Instance network_info: |[{"id": "7bb00d79-430f-44e2-9e92-57109c05d0bb", "address": "fa:16:3e:94:48:0c", "network": {"id": "e2ad0587-12a2-40ce-b2d0-c682042335fb", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1549110976-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1845e17fb7e64cf594c4e81c67ad8243", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c405e9f-a6c8-4308-acac-071654efe18e", "external-id": "nsx-vlan-transportzone-851", "segmentation_id": 851, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7bb00d79-43", "ovs_interfaceid": "7bb00d79-430f-44e2-9e92-57109c05d0bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 913.132380] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:94:48:0c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3c405e9f-a6c8-4308-acac-071654efe18e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7bb00d79-430f-44e2-9e92-57109c05d0bb', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 913.140029] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Creating folder: Project (1845e17fb7e64cf594c4e81c67ad8243). Parent ref: group-v281134. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 913.140335] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a65d556c-4a02-449c-be80-efa857a34804 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.162466] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Created folder: Project (1845e17fb7e64cf594c4e81c67ad8243) in parent group-v281134. [ 913.162683] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Creating folder: Instances. Parent ref: group-v281236. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 913.162940] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0b62f0fe-12dc-4d1d-a6d7-d5dad465d6b5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.176670] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Created folder: Instances in parent group-v281236. [ 913.176824] env[62460]: DEBUG oslo.service.loopingcall [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 913.177311] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-48989aff-581d-4da7-8344-7d4e8beb1358 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 913.177563] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 913.177775] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-55e8b6f3-0c9f-426d-a0b6-f4d027eafe98 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.182056] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-123880a3-5196-4c2c-bc47-a918ac334940 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.203435] env[62460]: DEBUG oslo_vmware.api [None req-48989aff-581d-4da7-8344-7d4e8beb1358 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 913.203435] env[62460]: value = "task-1313827" [ 913.203435] env[62460]: _type = "Task" [ 913.203435] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.203692] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 913.203692] env[62460]: value = "task-1313828" [ 913.203692] env[62460]: _type = "Task" [ 913.203692] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.221209] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313828, 'name': CreateVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.223873] env[62460]: DEBUG oslo_vmware.api [None req-48989aff-581d-4da7-8344-7d4e8beb1358 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313827, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.262583] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf098eb3-adf4-48c2-99ff-b20af9f5fd1f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.270121] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4dec8f4-ea3b-4131-9ce2-3c106eb00c55 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.302504] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48247784-2aad-42b8-8ebf-506444aa28f3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.306462] env[62460]: DEBUG nova.compute.manager [req-af72ef56-8de1-4567-bf61-2ca7a6d68601 req-0953f107-2f15-4512-9a07-10f39ef38968 service nova] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Received event network-changed-7bb00d79-430f-44e2-9e92-57109c05d0bb {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 913.306681] env[62460]: DEBUG nova.compute.manager [req-af72ef56-8de1-4567-bf61-2ca7a6d68601 req-0953f107-2f15-4512-9a07-10f39ef38968 service nova] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Refreshing instance network info cache due to event network-changed-7bb00d79-430f-44e2-9e92-57109c05d0bb. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 913.306894] env[62460]: DEBUG oslo_concurrency.lockutils [req-af72ef56-8de1-4567-bf61-2ca7a6d68601 req-0953f107-2f15-4512-9a07-10f39ef38968 service nova] Acquiring lock "refresh_cache-bf28fc98-d9a9-4678-a107-ace2b6503353" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.307056] env[62460]: DEBUG oslo_concurrency.lockutils [req-af72ef56-8de1-4567-bf61-2ca7a6d68601 req-0953f107-2f15-4512-9a07-10f39ef38968 service nova] Acquired lock "refresh_cache-bf28fc98-d9a9-4678-a107-ace2b6503353" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.307226] env[62460]: DEBUG nova.network.neutron [req-af72ef56-8de1-4567-bf61-2ca7a6d68601 req-0953f107-2f15-4512-9a07-10f39ef38968 service nova] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Refreshing network info cache for port 7bb00d79-430f-44e2-9e92-57109c05d0bb {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 913.316665] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83f9acd2-37ee-4d3e-8385-7727b16e97f2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.338019] env[62460]: DEBUG nova.compute.provider_tree [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 913.411129] env[62460]: DEBUG oslo_vmware.api [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313824, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.599453] env[62460]: DEBUG oslo_vmware.api [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313822, 'name': Destroy_Task, 'duration_secs': 1.017186} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.599879] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Destroyed the VM [ 913.599990] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Deleting Snapshot of the VM instance {{(pid=62460) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 913.600292] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-caf20fa5-d932-4917-9862-d34ab3acc156 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.606919] env[62460]: DEBUG oslo_vmware.api [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 913.606919] env[62460]: value = "task-1313829" [ 913.606919] env[62460]: _type = "Task" [ 913.606919] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.615319] env[62460]: DEBUG oslo_vmware.api [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313829, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.719955] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313828, 'name': CreateVM_Task} progress is 99%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.723436] env[62460]: DEBUG oslo_vmware.api [None req-48989aff-581d-4da7-8344-7d4e8beb1358 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313827, 'name': PowerOffVM_Task, 'duration_secs': 0.343413} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.723747] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-48989aff-581d-4da7-8344-7d4e8beb1358 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 913.723969] env[62460]: DEBUG nova.compute.manager [None req-48989aff-581d-4da7-8344-7d4e8beb1358 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 913.724868] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-421aeb4e-2ed5-42e1-b1a1-86b852eda953 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.844267] env[62460]: DEBUG nova.scheduler.client.report [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 913.912355] env[62460]: DEBUG oslo_vmware.api [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313824, 'name': ReconfigVM_Task, 'duration_secs': 0.845673} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.912701] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Reconfigured VM instance instance-00000055 to attach disk [datastore2] 6b084ed4-e8f0-4063-bf2a-252740640753/6b084ed4-e8f0-4063-bf2a-252740640753.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 913.913634] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-176e0922-64b4-49e5-829e-f1b1a71b4d7c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.920185] env[62460]: DEBUG oslo_vmware.api [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 913.920185] env[62460]: value = "task-1313830" [ 913.920185] env[62460]: _type = "Task" [ 913.920185] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.928421] env[62460]: DEBUG oslo_vmware.api [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313830, 'name': Rename_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.039393] env[62460]: DEBUG nova.network.neutron [req-af72ef56-8de1-4567-bf61-2ca7a6d68601 req-0953f107-2f15-4512-9a07-10f39ef38968 service nova] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Updated VIF entry in instance network info cache for port 7bb00d79-430f-44e2-9e92-57109c05d0bb. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 914.039790] env[62460]: DEBUG nova.network.neutron [req-af72ef56-8de1-4567-bf61-2ca7a6d68601 req-0953f107-2f15-4512-9a07-10f39ef38968 service nova] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Updating instance_info_cache with network_info: [{"id": "7bb00d79-430f-44e2-9e92-57109c05d0bb", "address": "fa:16:3e:94:48:0c", "network": {"id": "e2ad0587-12a2-40ce-b2d0-c682042335fb", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1549110976-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1845e17fb7e64cf594c4e81c67ad8243", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c405e9f-a6c8-4308-acac-071654efe18e", "external-id": "nsx-vlan-transportzone-851", "segmentation_id": 851, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7bb00d79-43", "ovs_interfaceid": "7bb00d79-430f-44e2-9e92-57109c05d0bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.117898] env[62460]: DEBUG oslo_vmware.api [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313829, 'name': RemoveSnapshot_Task, 'duration_secs': 0.352649} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.118207] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Deleted Snapshot of the VM instance {{(pid=62460) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 914.118490] env[62460]: DEBUG nova.compute.manager [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 914.119360] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0731ce63-f481-4781-a069-9ce83eed4d54 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.215513] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313828, 'name': CreateVM_Task, 'duration_secs': 0.565602} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.215703] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 914.216319] env[62460]: DEBUG oslo_concurrency.lockutils [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.216497] env[62460]: DEBUG oslo_concurrency.lockutils [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.216832] env[62460]: DEBUG oslo_concurrency.lockutils [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 914.217099] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-04b2713f-6b63-4e7b-9469-44435e40218d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.221478] env[62460]: DEBUG oslo_vmware.api [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Waiting for the task: (returnval){ [ 914.221478] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]5274d7f3-33c7-fb22-e0cc-9600ae5538b0" [ 914.221478] env[62460]: _type = "Task" [ 914.221478] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.229031] env[62460]: DEBUG oslo_vmware.api [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5274d7f3-33c7-fb22-e0cc-9600ae5538b0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.236560] env[62460]: DEBUG oslo_concurrency.lockutils [None req-48989aff-581d-4da7-8344-7d4e8beb1358 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "09155629-51c8-4043-b1ee-6d5036552a67" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.573s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.349403] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.350s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.349975] env[62460]: DEBUG nova.compute.manager [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 914.354039] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a804a375-a4e7-4118-8997-76c65de0dcd0 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.257s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.354191] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a804a375-a4e7-4118-8997-76c65de0dcd0 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.356072] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b138fff5-d93a-4df7-93dc-d05c5fa66426 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.194s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.356270] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b138fff5-d93a-4df7-93dc-d05c5fa66426 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.358021] env[62460]: DEBUG oslo_concurrency.lockutils [None req-8f649042-5fd7-4f94-80db-d6a078b9ce5a tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.118s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.358249] env[62460]: DEBUG oslo_concurrency.lockutils [None req-8f649042-5fd7-4f94-80db-d6a078b9ce5a tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.359795] env[62460]: DEBUG oslo_concurrency.lockutils [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.120s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.361209] env[62460]: INFO nova.compute.claims [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 914.383160] env[62460]: INFO nova.scheduler.client.report [None req-b138fff5-d93a-4df7-93dc-d05c5fa66426 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Deleted allocations for instance 0da70720-197c-4483-93e1-019e0f040c4c [ 914.384920] env[62460]: INFO nova.scheduler.client.report [None req-8f649042-5fd7-4f94-80db-d6a078b9ce5a tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Deleted allocations for instance 8bcaa672-97c8-466a-812c-5723e08d0c52 [ 914.396194] env[62460]: INFO nova.scheduler.client.report [None req-a804a375-a4e7-4118-8997-76c65de0dcd0 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Deleted allocations for instance fbe558f1-ce5a-4429-8833-48bdabe91071 [ 914.429401] env[62460]: DEBUG oslo_vmware.api [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313830, 'name': Rename_Task, 'duration_secs': 0.124406} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.429715] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 914.429968] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d3f1e529-090e-4039-8335-4e8febde5611 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.436768] env[62460]: DEBUG oslo_vmware.api [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 914.436768] env[62460]: value = "task-1313831" [ 914.436768] env[62460]: _type = "Task" [ 914.436768] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.444802] env[62460]: DEBUG oslo_vmware.api [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313831, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.542665] env[62460]: DEBUG oslo_concurrency.lockutils [req-af72ef56-8de1-4567-bf61-2ca7a6d68601 req-0953f107-2f15-4512-9a07-10f39ef38968 service nova] Releasing lock "refresh_cache-bf28fc98-d9a9-4678-a107-ace2b6503353" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.634989] env[62460]: INFO nova.compute.manager [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Shelve offloading [ 914.639836] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 914.640132] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-54c08670-532e-424f-93be-8844b140a775 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.647874] env[62460]: DEBUG oslo_vmware.api [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 914.647874] env[62460]: value = "task-1313832" [ 914.647874] env[62460]: _type = "Task" [ 914.647874] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.657666] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] VM already powered off {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 914.657901] env[62460]: DEBUG nova.compute.manager [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 914.658674] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e10e6fff-1043-437f-bd05-142f8708d43f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.664643] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquiring lock "refresh_cache-896f8ff4-e45f-4403-a727-03ee25e58609" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.664823] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquired lock "refresh_cache-896f8ff4-e45f-4403-a727-03ee25e58609" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.665014] env[62460]: DEBUG nova.network.neutron [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 914.731986] env[62460]: DEBUG oslo_vmware.api [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5274d7f3-33c7-fb22-e0cc-9600ae5538b0, 'name': SearchDatastore_Task, 'duration_secs': 0.008821} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.732345] env[62460]: DEBUG oslo_concurrency.lockutils [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.732624] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 914.732821] env[62460]: DEBUG oslo_concurrency.lockutils [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.732977] env[62460]: DEBUG oslo_concurrency.lockutils [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.733180] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 914.733559] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ace9c46f-a049-4890-8e44-6091dc4689df {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.742173] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 914.742381] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 914.743126] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2e501778-8061-4ad3-bf2f-8cf65ef67e2b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.748422] env[62460]: DEBUG oslo_vmware.api [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Waiting for the task: (returnval){ [ 914.748422] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52c82652-c9d2-9b7f-83a4-edd9c5d2d31f" [ 914.748422] env[62460]: _type = "Task" [ 914.748422] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.756706] env[62460]: DEBUG oslo_vmware.api [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52c82652-c9d2-9b7f-83a4-edd9c5d2d31f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.855552] env[62460]: DEBUG nova.compute.utils [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 914.857337] env[62460]: DEBUG nova.compute.manager [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 914.857545] env[62460]: DEBUG nova.network.neutron [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 914.896031] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b138fff5-d93a-4df7-93dc-d05c5fa66426 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Lock "0da70720-197c-4483-93e1-019e0f040c4c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.843s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.897626] env[62460]: DEBUG oslo_concurrency.lockutils [None req-8f649042-5fd7-4f94-80db-d6a078b9ce5a tempest-InstanceActionsV221TestJSON-2053132648 tempest-InstanceActionsV221TestJSON-2053132648-project-member] Lock "8bcaa672-97c8-466a-812c-5723e08d0c52" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.541s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.903356] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a804a375-a4e7-4118-8997-76c65de0dcd0 tempest-MultipleCreateTestJSON-214463140 tempest-MultipleCreateTestJSON-214463140-project-member] Lock "fbe558f1-ce5a-4429-8833-48bdabe91071" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.774s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.907987] env[62460]: DEBUG nova.policy [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd7b38fb0a53f4d078ae86d68e3eb6268', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '53fa06008e9f43488362895e7a143700', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 914.947974] env[62460]: DEBUG oslo_vmware.api [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313831, 'name': PowerOnVM_Task, 'duration_secs': 0.447192} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.952021] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 914.952021] env[62460]: INFO nova.compute.manager [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Took 9.34 seconds to spawn the instance on the hypervisor. [ 914.952021] env[62460]: DEBUG nova.compute.manager [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 914.952021] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34f15456-9d25-4d33-b2fe-adbc601a0692 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.140711] env[62460]: DEBUG oslo_concurrency.lockutils [None req-dc607155-ae9c-45e1-8539-0d765fe9cad5 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "09155629-51c8-4043-b1ee-6d5036552a67" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.141438] env[62460]: DEBUG oslo_concurrency.lockutils [None req-dc607155-ae9c-45e1-8539-0d765fe9cad5 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "09155629-51c8-4043-b1ee-6d5036552a67" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.144658] env[62460]: DEBUG oslo_concurrency.lockutils [None req-dc607155-ae9c-45e1-8539-0d765fe9cad5 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "09155629-51c8-4043-b1ee-6d5036552a67-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.144658] env[62460]: DEBUG oslo_concurrency.lockutils [None req-dc607155-ae9c-45e1-8539-0d765fe9cad5 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "09155629-51c8-4043-b1ee-6d5036552a67-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.144658] env[62460]: DEBUG oslo_concurrency.lockutils [None req-dc607155-ae9c-45e1-8539-0d765fe9cad5 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "09155629-51c8-4043-b1ee-6d5036552a67-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.146766] env[62460]: INFO nova.compute.manager [None req-dc607155-ae9c-45e1-8539-0d765fe9cad5 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Terminating instance [ 915.153018] env[62460]: DEBUG nova.compute.manager [None req-dc607155-ae9c-45e1-8539-0d765fe9cad5 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 915.153018] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-dc607155-ae9c-45e1-8539-0d765fe9cad5 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 915.153018] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7565e28b-3b9f-44c4-a8f5-d52a975bf591 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.159743] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-dc607155-ae9c-45e1-8539-0d765fe9cad5 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 915.159996] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-39659ebf-4253-4649-9f10-05c80d95b5e3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.231477] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-dc607155-ae9c-45e1-8539-0d765fe9cad5 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 915.231780] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-dc607155-ae9c-45e1-8539-0d765fe9cad5 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Deleting contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 915.231911] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc607155-ae9c-45e1-8539-0d765fe9cad5 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Deleting the datastore file [datastore2] 09155629-51c8-4043-b1ee-6d5036552a67 {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 915.232286] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0c34b5a6-e402-44fe-b582-be6f0c181e18 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.240154] env[62460]: DEBUG oslo_vmware.api [None req-dc607155-ae9c-45e1-8539-0d765fe9cad5 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 915.240154] env[62460]: value = "task-1313834" [ 915.240154] env[62460]: _type = "Task" [ 915.240154] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.253466] env[62460]: DEBUG oslo_vmware.api [None req-dc607155-ae9c-45e1-8539-0d765fe9cad5 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313834, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.264662] env[62460]: DEBUG oslo_vmware.api [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52c82652-c9d2-9b7f-83a4-edd9c5d2d31f, 'name': SearchDatastore_Task, 'duration_secs': 0.008343} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.265539] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd0d6bdd-3978-4bc2-9733-f451e50387d6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.271331] env[62460]: DEBUG oslo_vmware.api [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Waiting for the task: (returnval){ [ 915.271331] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]524743db-ec96-62fa-5b75-bf3158ef877f" [ 915.271331] env[62460]: _type = "Task" [ 915.271331] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.281224] env[62460]: DEBUG oslo_vmware.api [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]524743db-ec96-62fa-5b75-bf3158ef877f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.331954] env[62460]: DEBUG nova.network.neutron [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Successfully created port: 546b0347-51db-4a4a-99e9-4f3db96a36bd {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 915.361113] env[62460]: DEBUG nova.compute.manager [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 915.471846] env[62460]: INFO nova.compute.manager [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Took 24.68 seconds to build instance. [ 915.543829] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca49f73a-2729-483e-a91b-33a9790a98f2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.555119] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4be2027-bdfa-4ba5-bd8b-4bab3e5a0ece {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.589473] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8293ce50-2004-48a1-9d39-6086a889c418 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.598146] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09172901-48a3-455d-99d5-cf4bb239d972 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.613831] env[62460]: DEBUG nova.compute.provider_tree [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 915.751792] env[62460]: DEBUG oslo_vmware.api [None req-dc607155-ae9c-45e1-8539-0d765fe9cad5 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313834, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.153779} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.752432] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc607155-ae9c-45e1-8539-0d765fe9cad5 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 915.752634] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-dc607155-ae9c-45e1-8539-0d765fe9cad5 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Deleted contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 915.752823] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-dc607155-ae9c-45e1-8539-0d765fe9cad5 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 915.753128] env[62460]: INFO nova.compute.manager [None req-dc607155-ae9c-45e1-8539-0d765fe9cad5 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Took 0.60 seconds to destroy the instance on the hypervisor. [ 915.753367] env[62460]: DEBUG oslo.service.loopingcall [None req-dc607155-ae9c-45e1-8539-0d765fe9cad5 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 915.753498] env[62460]: DEBUG nova.compute.manager [-] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 915.753608] env[62460]: DEBUG nova.network.neutron [-] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 915.770536] env[62460]: DEBUG nova.network.neutron [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Updating instance_info_cache with network_info: [{"id": "72a87ddf-0585-429a-b9de-d73bcad42cd1", "address": "fa:16:3e:85:2f:50", "network": {"id": "b5cb4ee0-4562-45be-a25d-51c3a5932efd", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-823340443-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.149", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3d2d1c48ec14121a2e8c9b3f800a949", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55c757ac-f8b2-466d-b634-07dbd100b312", "external-id": "nsx-vlan-transportzone-159", "segmentation_id": 159, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap72a87ddf-05", "ovs_interfaceid": "72a87ddf-0585-429a-b9de-d73bcad42cd1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.787557] env[62460]: DEBUG oslo_vmware.api [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]524743db-ec96-62fa-5b75-bf3158ef877f, 'name': SearchDatastore_Task, 'duration_secs': 0.020236} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.788851] env[62460]: DEBUG oslo_concurrency.lockutils [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.788851] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] bf28fc98-d9a9-4678-a107-ace2b6503353/bf28fc98-d9a9-4678-a107-ace2b6503353.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 915.789018] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-72dc5ee4-6cdf-4b90-8bbf-ea28cbd86cb4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.798205] env[62460]: DEBUG oslo_vmware.api [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Waiting for the task: (returnval){ [ 915.798205] env[62460]: value = "task-1313835" [ 915.798205] env[62460]: _type = "Task" [ 915.798205] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.810393] env[62460]: DEBUG oslo_vmware.api [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Task: {'id': task-1313835, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.973806] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c3e6f49b-f2b5-4f36-aa25-c45bd95e8b06 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "6b084ed4-e8f0-4063-bf2a-252740640753" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.188s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.036571] env[62460]: DEBUG oslo_concurrency.lockutils [None req-bd5686d8-0c05-4196-a00b-777bd0158d91 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Acquiring lock "3e10af70-db52-4d4a-bb92-821a05dcbab6" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.036932] env[62460]: DEBUG oslo_concurrency.lockutils [None req-bd5686d8-0c05-4196-a00b-777bd0158d91 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Lock "3e10af70-db52-4d4a-bb92-821a05dcbab6" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.118414] env[62460]: DEBUG nova.scheduler.client.report [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 916.279489] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Releasing lock "refresh_cache-896f8ff4-e45f-4403-a727-03ee25e58609" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.310794] env[62460]: DEBUG oslo_vmware.api [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Task: {'id': task-1313835, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.481134} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.311074] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] bf28fc98-d9a9-4678-a107-ace2b6503353/bf28fc98-d9a9-4678-a107-ace2b6503353.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 916.311300] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 916.311553] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-12ac47a6-cc77-448f-8ff1-b2abaedf4e05 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.318547] env[62460]: DEBUG oslo_vmware.api [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Waiting for the task: (returnval){ [ 916.318547] env[62460]: value = "task-1313836" [ 916.318547] env[62460]: _type = "Task" [ 916.318547] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.330556] env[62460]: DEBUG oslo_vmware.api [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Task: {'id': task-1313836, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.372088] env[62460]: DEBUG nova.compute.manager [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 916.406472] env[62460]: DEBUG nova.virt.hardware [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 916.406573] env[62460]: DEBUG nova.virt.hardware [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 916.406757] env[62460]: DEBUG nova.virt.hardware [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 916.406874] env[62460]: DEBUG nova.virt.hardware [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 916.407675] env[62460]: DEBUG nova.virt.hardware [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 916.407938] env[62460]: DEBUG nova.virt.hardware [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 916.408198] env[62460]: DEBUG nova.virt.hardware [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 916.408376] env[62460]: DEBUG nova.virt.hardware [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 916.408553] env[62460]: DEBUG nova.virt.hardware [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 916.408755] env[62460]: DEBUG nova.virt.hardware [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 916.408950] env[62460]: DEBUG nova.virt.hardware [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 916.409930] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e61fc914-aab8-47df-8a8e-0adbdf497090 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.417921] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc0d316a-3d2c-4bd3-bb78-cd76d446dd40 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.503667] env[62460]: DEBUG nova.compute.manager [req-94171fab-17aa-4fd6-ab12-2cf3875da099 req-7f5918cf-9375-425e-9f8b-b65bbf3cf91a service nova] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Received event network-vif-deleted-dbd267c2-b2d8-46ac-8d82-f645dd8560d5 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 916.503667] env[62460]: INFO nova.compute.manager [req-94171fab-17aa-4fd6-ab12-2cf3875da099 req-7f5918cf-9375-425e-9f8b-b65bbf3cf91a service nova] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Neutron deleted interface dbd267c2-b2d8-46ac-8d82-f645dd8560d5; detaching it from the instance and deleting it from the info cache [ 916.503667] env[62460]: DEBUG nova.network.neutron [req-94171fab-17aa-4fd6-ab12-2cf3875da099 req-7f5918cf-9375-425e-9f8b-b65bbf3cf91a service nova] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.541098] env[62460]: DEBUG nova.compute.utils [None req-bd5686d8-0c05-4196-a00b-777bd0158d91 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 916.633330] env[62460]: DEBUG oslo_concurrency.lockutils [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.273s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.633882] env[62460]: DEBUG nova.compute.manager [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 916.637598] env[62460]: DEBUG oslo_concurrency.lockutils [None req-07205f12-9c2f-4768-9b9d-0f27bbfe3d9c tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.374s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.637842] env[62460]: DEBUG oslo_concurrency.lockutils [None req-07205f12-9c2f-4768-9b9d-0f27bbfe3d9c tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.663144] env[62460]: DEBUG nova.network.neutron [-] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.671629] env[62460]: INFO nova.scheduler.client.report [None req-07205f12-9c2f-4768-9b9d-0f27bbfe3d9c tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Deleted allocations for instance f1921907-f69c-49a3-896c-a4b12b526ffa [ 916.815296] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 916.816405] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7013cd27-9068-475f-9a1a-8d81d9e817e9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.830678] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 916.832140] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1da33a62-cd2b-4614-ae19-c75e60f47edb {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.833587] env[62460]: DEBUG oslo_vmware.api [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Task: {'id': task-1313836, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063084} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.833853] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 916.835033] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-730083f5-2e07-4586-ab94-9f254fa587e9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.856972] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Reconfiguring VM instance instance-00000056 to attach disk [datastore2] bf28fc98-d9a9-4678-a107-ace2b6503353/bf28fc98-d9a9-4678-a107-ace2b6503353.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 916.857683] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-284830f6-5e72-4294-b775-9fd321f2be9f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.877685] env[62460]: DEBUG oslo_vmware.api [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Waiting for the task: (returnval){ [ 916.877685] env[62460]: value = "task-1313838" [ 916.877685] env[62460]: _type = "Task" [ 916.877685] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.886027] env[62460]: DEBUG oslo_vmware.api [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Task: {'id': task-1313838, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.902094] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 916.902339] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Deleting contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 916.902524] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Deleting the datastore file [datastore1] 896f8ff4-e45f-4403-a727-03ee25e58609 {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 916.902793] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0c65075c-9d3f-478c-8e51-4064bf234bba {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.908925] env[62460]: DEBUG oslo_vmware.api [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 916.908925] env[62460]: value = "task-1313839" [ 916.908925] env[62460]: _type = "Task" [ 916.908925] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.917879] env[62460]: DEBUG oslo_vmware.api [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313839, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.006886] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7fa313c3-e112-45f1-b6e5-ba2b2f7b1b1b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.018333] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f245f5c0-f8f4-4df2-b606-8d1e32098875 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.047653] env[62460]: DEBUG oslo_concurrency.lockutils [None req-bd5686d8-0c05-4196-a00b-777bd0158d91 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Lock "3e10af70-db52-4d4a-bb92-821a05dcbab6" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.011s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.048441] env[62460]: DEBUG nova.compute.manager [req-94171fab-17aa-4fd6-ab12-2cf3875da099 req-7f5918cf-9375-425e-9f8b-b65bbf3cf91a service nova] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Detach interface failed, port_id=dbd267c2-b2d8-46ac-8d82-f645dd8560d5, reason: Instance 09155629-51c8-4043-b1ee-6d5036552a67 could not be found. {{(pid=62460) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 917.139809] env[62460]: DEBUG nova.compute.utils [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 917.142350] env[62460]: DEBUG nova.compute.manager [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 917.142350] env[62460]: DEBUG nova.network.neutron [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 917.166127] env[62460]: INFO nova.compute.manager [-] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Took 1.41 seconds to deallocate network for instance. [ 917.183902] env[62460]: DEBUG oslo_concurrency.lockutils [None req-07205f12-9c2f-4768-9b9d-0f27bbfe3d9c tempest-ServersTestBootFromVolume-866305222 tempest-ServersTestBootFromVolume-866305222-project-member] Lock "f1921907-f69c-49a3-896c-a4b12b526ffa" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.089s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.231813] env[62460]: DEBUG nova.policy [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '632d7ca8cb4243f990d741a3f70f422a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '10ac54db1ef54e249a077bbb3dde4242', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 917.390806] env[62460]: DEBUG oslo_vmware.api [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Task: {'id': task-1313838, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.420756] env[62460]: DEBUG oslo_vmware.api [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313839, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.168051} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.421137] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 917.421427] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Deleted contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 917.421708] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 917.430535] env[62460]: DEBUG nova.network.neutron [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Successfully updated port: 546b0347-51db-4a4a-99e9-4f3db96a36bd {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 917.455231] env[62460]: INFO nova.scheduler.client.report [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Deleted allocations for instance 896f8ff4-e45f-4403-a727-03ee25e58609 [ 917.630476] env[62460]: DEBUG nova.compute.manager [req-8ac26b0b-b135-4f64-847a-bfdc6f3b2601 req-89c89c31-7d82-4a3b-8735-124d75906447 service nova] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Received event network-vif-plugged-546b0347-51db-4a4a-99e9-4f3db96a36bd {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 917.630476] env[62460]: DEBUG oslo_concurrency.lockutils [req-8ac26b0b-b135-4f64-847a-bfdc6f3b2601 req-89c89c31-7d82-4a3b-8735-124d75906447 service nova] Acquiring lock "abd4501b-3ad8-4103-921d-5b80fc2f313f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.630637] env[62460]: DEBUG oslo_concurrency.lockutils [req-8ac26b0b-b135-4f64-847a-bfdc6f3b2601 req-89c89c31-7d82-4a3b-8735-124d75906447 service nova] Lock "abd4501b-3ad8-4103-921d-5b80fc2f313f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.630781] env[62460]: DEBUG oslo_concurrency.lockutils [req-8ac26b0b-b135-4f64-847a-bfdc6f3b2601 req-89c89c31-7d82-4a3b-8735-124d75906447 service nova] Lock "abd4501b-3ad8-4103-921d-5b80fc2f313f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.630979] env[62460]: DEBUG nova.compute.manager [req-8ac26b0b-b135-4f64-847a-bfdc6f3b2601 req-89c89c31-7d82-4a3b-8735-124d75906447 service nova] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] No waiting events found dispatching network-vif-plugged-546b0347-51db-4a4a-99e9-4f3db96a36bd {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 917.631975] env[62460]: WARNING nova.compute.manager [req-8ac26b0b-b135-4f64-847a-bfdc6f3b2601 req-89c89c31-7d82-4a3b-8735-124d75906447 service nova] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Received unexpected event network-vif-plugged-546b0347-51db-4a4a-99e9-4f3db96a36bd for instance with vm_state building and task_state spawning. [ 917.641881] env[62460]: DEBUG nova.network.neutron [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Successfully created port: 3304dbf2-549a-48d4-bd32-9fbd3d98cc3d {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 917.646233] env[62460]: DEBUG nova.compute.manager [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 917.683544] env[62460]: DEBUG oslo_concurrency.lockutils [None req-dc607155-ae9c-45e1-8539-0d765fe9cad5 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.684383] env[62460]: DEBUG oslo_concurrency.lockutils [None req-dc607155-ae9c-45e1-8539-0d765fe9cad5 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.684383] env[62460]: DEBUG nova.objects.instance [None req-dc607155-ae9c-45e1-8539-0d765fe9cad5 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lazy-loading 'resources' on Instance uuid 09155629-51c8-4043-b1ee-6d5036552a67 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 917.889134] env[62460]: DEBUG oslo_vmware.api [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Task: {'id': task-1313838, 'name': ReconfigVM_Task, 'duration_secs': 0.611254} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.889552] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Reconfigured VM instance instance-00000056 to attach disk [datastore2] bf28fc98-d9a9-4678-a107-ace2b6503353/bf28fc98-d9a9-4678-a107-ace2b6503353.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 917.890272] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-08e55e67-7d3c-4b9e-bc62-c87bc9e8b803 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.897033] env[62460]: DEBUG oslo_vmware.api [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Waiting for the task: (returnval){ [ 917.897033] env[62460]: value = "task-1313840" [ 917.897033] env[62460]: _type = "Task" [ 917.897033] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.905616] env[62460]: DEBUG oslo_vmware.api [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Task: {'id': task-1313840, 'name': Rename_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.933856] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "refresh_cache-abd4501b-3ad8-4103-921d-5b80fc2f313f" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.934112] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquired lock "refresh_cache-abd4501b-3ad8-4103-921d-5b80fc2f313f" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.934286] env[62460]: DEBUG nova.network.neutron [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 917.960605] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.134267] env[62460]: DEBUG oslo_concurrency.lockutils [None req-bd5686d8-0c05-4196-a00b-777bd0158d91 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Acquiring lock "3e10af70-db52-4d4a-bb92-821a05dcbab6" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.135054] env[62460]: DEBUG oslo_concurrency.lockutils [None req-bd5686d8-0c05-4196-a00b-777bd0158d91 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Lock "3e10af70-db52-4d4a-bb92-821a05dcbab6" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.135333] env[62460]: INFO nova.compute.manager [None req-bd5686d8-0c05-4196-a00b-777bd0158d91 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Attaching volume 8131193e-aa4f-424e-b51d-e2276d84089e to /dev/sdb [ 918.177887] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ad76a6e-89b4-4a33-a1d5-48dc43454349 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.187816] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b456cd22-607d-44d1-8c1e-336938b9cc8e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.210789] env[62460]: DEBUG nova.virt.block_device [None req-bd5686d8-0c05-4196-a00b-777bd0158d91 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Updating existing volume attachment record: 3dc81190-8ebb-4440-a43c-da8c6668d405 {{(pid=62460) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 918.352262] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67eac2b6-808d-4c1e-b26e-11506df860d4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.362015] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97cbfc59-28f5-4efe-8267-7069d4d94c67 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.396452] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e4b82e0-d1eb-4212-8287-bd90ddd4a7e8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.409109] env[62460]: DEBUG oslo_vmware.api [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Task: {'id': task-1313840, 'name': Rename_Task, 'duration_secs': 0.317741} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.411173] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 918.411505] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bf081e0f-ede0-4dcf-a36a-72f185d69f85 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.414857] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63494e21-b8b9-4021-a79a-7fc840196d90 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.431401] env[62460]: DEBUG nova.compute.provider_tree [None req-dc607155-ae9c-45e1-8539-0d765fe9cad5 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 918.436999] env[62460]: DEBUG oslo_vmware.api [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Waiting for the task: (returnval){ [ 918.436999] env[62460]: value = "task-1313842" [ 918.436999] env[62460]: _type = "Task" [ 918.436999] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.450616] env[62460]: DEBUG oslo_vmware.api [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Task: {'id': task-1313842, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.493058] env[62460]: DEBUG nova.network.neutron [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 918.541456] env[62460]: DEBUG nova.compute.manager [req-0f11ea75-6eb2-433c-ba70-dcdf588014ad req-845320bf-d1b0-4c53-87e5-4b5933eb6209 service nova] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Received event network-vif-unplugged-72a87ddf-0585-429a-b9de-d73bcad42cd1 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 918.541687] env[62460]: DEBUG oslo_concurrency.lockutils [req-0f11ea75-6eb2-433c-ba70-dcdf588014ad req-845320bf-d1b0-4c53-87e5-4b5933eb6209 service nova] Acquiring lock "896f8ff4-e45f-4403-a727-03ee25e58609-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.541884] env[62460]: DEBUG oslo_concurrency.lockutils [req-0f11ea75-6eb2-433c-ba70-dcdf588014ad req-845320bf-d1b0-4c53-87e5-4b5933eb6209 service nova] Lock "896f8ff4-e45f-4403-a727-03ee25e58609-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.542070] env[62460]: DEBUG oslo_concurrency.lockutils [req-0f11ea75-6eb2-433c-ba70-dcdf588014ad req-845320bf-d1b0-4c53-87e5-4b5933eb6209 service nova] Lock "896f8ff4-e45f-4403-a727-03ee25e58609-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.542292] env[62460]: DEBUG nova.compute.manager [req-0f11ea75-6eb2-433c-ba70-dcdf588014ad req-845320bf-d1b0-4c53-87e5-4b5933eb6209 service nova] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] No waiting events found dispatching network-vif-unplugged-72a87ddf-0585-429a-b9de-d73bcad42cd1 {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 918.542475] env[62460]: WARNING nova.compute.manager [req-0f11ea75-6eb2-433c-ba70-dcdf588014ad req-845320bf-d1b0-4c53-87e5-4b5933eb6209 service nova] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Received unexpected event network-vif-unplugged-72a87ddf-0585-429a-b9de-d73bcad42cd1 for instance with vm_state shelved_offloaded and task_state None. [ 918.542648] env[62460]: DEBUG nova.compute.manager [req-0f11ea75-6eb2-433c-ba70-dcdf588014ad req-845320bf-d1b0-4c53-87e5-4b5933eb6209 service nova] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Received event network-changed-72a87ddf-0585-429a-b9de-d73bcad42cd1 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 918.542827] env[62460]: DEBUG nova.compute.manager [req-0f11ea75-6eb2-433c-ba70-dcdf588014ad req-845320bf-d1b0-4c53-87e5-4b5933eb6209 service nova] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Refreshing instance network info cache due to event network-changed-72a87ddf-0585-429a-b9de-d73bcad42cd1. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 918.543010] env[62460]: DEBUG oslo_concurrency.lockutils [req-0f11ea75-6eb2-433c-ba70-dcdf588014ad req-845320bf-d1b0-4c53-87e5-4b5933eb6209 service nova] Acquiring lock "refresh_cache-896f8ff4-e45f-4403-a727-03ee25e58609" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.543164] env[62460]: DEBUG oslo_concurrency.lockutils [req-0f11ea75-6eb2-433c-ba70-dcdf588014ad req-845320bf-d1b0-4c53-87e5-4b5933eb6209 service nova] Acquired lock "refresh_cache-896f8ff4-e45f-4403-a727-03ee25e58609" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.543328] env[62460]: DEBUG nova.network.neutron [req-0f11ea75-6eb2-433c-ba70-dcdf588014ad req-845320bf-d1b0-4c53-87e5-4b5933eb6209 service nova] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Refreshing network info cache for port 72a87ddf-0585-429a-b9de-d73bcad42cd1 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 918.659014] env[62460]: DEBUG nova.compute.manager [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 918.687521] env[62460]: DEBUG nova.virt.hardware [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 918.687804] env[62460]: DEBUG nova.virt.hardware [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 918.687975] env[62460]: DEBUG nova.virt.hardware [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 918.688588] env[62460]: DEBUG nova.virt.hardware [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 918.688826] env[62460]: DEBUG nova.virt.hardware [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 918.689029] env[62460]: DEBUG nova.virt.hardware [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 918.689295] env[62460]: DEBUG nova.virt.hardware [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 918.689500] env[62460]: DEBUG nova.virt.hardware [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 918.689771] env[62460]: DEBUG nova.virt.hardware [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 918.689981] env[62460]: DEBUG nova.virt.hardware [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 918.690227] env[62460]: DEBUG nova.virt.hardware [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 918.692440] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4713b5e5-28f5-4682-a2b9-5c2581ea10e4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.703829] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c24b9f94-a4c4-48e7-8d3e-3882a18c961b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.720438] env[62460]: DEBUG nova.network.neutron [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Updating instance_info_cache with network_info: [{"id": "546b0347-51db-4a4a-99e9-4f3db96a36bd", "address": "fa:16:3e:54:9b:7b", "network": {"id": "d3136e32-ad55-4b73-835c-8fa5f0480767", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1511414968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53fa06008e9f43488362895e7a143700", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "054fcd1e-638e-425a-a1de-78cb188ae026", "external-id": "nsx-vlan-transportzone-658", "segmentation_id": 658, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap546b0347-51", "ovs_interfaceid": "546b0347-51db-4a4a-99e9-4f3db96a36bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 918.938716] env[62460]: DEBUG nova.scheduler.client.report [None req-dc607155-ae9c-45e1-8539-0d765fe9cad5 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 918.953158] env[62460]: DEBUG oslo_vmware.api [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Task: {'id': task-1313842, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.227993] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Releasing lock "refresh_cache-abd4501b-3ad8-4103-921d-5b80fc2f313f" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.227993] env[62460]: DEBUG nova.compute.manager [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Instance network_info: |[{"id": "546b0347-51db-4a4a-99e9-4f3db96a36bd", "address": "fa:16:3e:54:9b:7b", "network": {"id": "d3136e32-ad55-4b73-835c-8fa5f0480767", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1511414968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53fa06008e9f43488362895e7a143700", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "054fcd1e-638e-425a-a1de-78cb188ae026", "external-id": "nsx-vlan-transportzone-658", "segmentation_id": 658, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap546b0347-51", "ovs_interfaceid": "546b0347-51db-4a4a-99e9-4f3db96a36bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 919.230046] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:54:9b:7b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '054fcd1e-638e-425a-a1de-78cb188ae026', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '546b0347-51db-4a4a-99e9-4f3db96a36bd', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 919.236553] env[62460]: DEBUG oslo.service.loopingcall [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 919.237381] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 919.237692] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b1a8bfae-b134-4eda-8d43-7427aa5f6afb {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.261765] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 919.261765] env[62460]: value = "task-1313845" [ 919.261765] env[62460]: _type = "Task" [ 919.261765] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.276647] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313845, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.279021] env[62460]: DEBUG nova.compute.manager [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Stashing vm_state: active {{(pid=62460) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 919.437047] env[62460]: DEBUG nova.network.neutron [req-0f11ea75-6eb2-433c-ba70-dcdf588014ad req-845320bf-d1b0-4c53-87e5-4b5933eb6209 service nova] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Updated VIF entry in instance network info cache for port 72a87ddf-0585-429a-b9de-d73bcad42cd1. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 919.437047] env[62460]: DEBUG nova.network.neutron [req-0f11ea75-6eb2-433c-ba70-dcdf588014ad req-845320bf-d1b0-4c53-87e5-4b5933eb6209 service nova] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Updating instance_info_cache with network_info: [{"id": "72a87ddf-0585-429a-b9de-d73bcad42cd1", "address": "fa:16:3e:85:2f:50", "network": {"id": "b5cb4ee0-4562-45be-a25d-51c3a5932efd", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-823340443-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.149", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3d2d1c48ec14121a2e8c9b3f800a949", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap72a87ddf-05", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.450356] env[62460]: DEBUG oslo_concurrency.lockutils [None req-dc607155-ae9c-45e1-8539-0d765fe9cad5 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.766s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.458430] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.498s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.459219] env[62460]: DEBUG nova.objects.instance [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lazy-loading 'resources' on Instance uuid 896f8ff4-e45f-4403-a727-03ee25e58609 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 919.461875] env[62460]: DEBUG oslo_vmware.api [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Task: {'id': task-1313842, 'name': PowerOnVM_Task, 'duration_secs': 0.534249} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.465571] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 919.465571] env[62460]: INFO nova.compute.manager [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Took 8.75 seconds to spawn the instance on the hypervisor. [ 919.465571] env[62460]: DEBUG nova.compute.manager [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 919.465571] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79cab050-8e08-4ca3-9410-f65040b0f856 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.480942] env[62460]: INFO nova.scheduler.client.report [None req-dc607155-ae9c-45e1-8539-0d765fe9cad5 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Deleted allocations for instance 09155629-51c8-4043-b1ee-6d5036552a67 [ 919.669588] env[62460]: DEBUG nova.compute.manager [req-ae4fcbd3-44d1-49be-ac2b-a55c0afd72a1 req-2a10b68e-9979-42c7-9f17-c3e332db911a service nova] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Received event network-changed-546b0347-51db-4a4a-99e9-4f3db96a36bd {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 919.669820] env[62460]: DEBUG nova.compute.manager [req-ae4fcbd3-44d1-49be-ac2b-a55c0afd72a1 req-2a10b68e-9979-42c7-9f17-c3e332db911a service nova] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Refreshing instance network info cache due to event network-changed-546b0347-51db-4a4a-99e9-4f3db96a36bd. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 919.670372] env[62460]: DEBUG oslo_concurrency.lockutils [req-ae4fcbd3-44d1-49be-ac2b-a55c0afd72a1 req-2a10b68e-9979-42c7-9f17-c3e332db911a service nova] Acquiring lock "refresh_cache-abd4501b-3ad8-4103-921d-5b80fc2f313f" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.674033] env[62460]: DEBUG oslo_concurrency.lockutils [req-ae4fcbd3-44d1-49be-ac2b-a55c0afd72a1 req-2a10b68e-9979-42c7-9f17-c3e332db911a service nova] Acquired lock "refresh_cache-abd4501b-3ad8-4103-921d-5b80fc2f313f" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.674033] env[62460]: DEBUG nova.network.neutron [req-ae4fcbd3-44d1-49be-ac2b-a55c0afd72a1 req-2a10b68e-9979-42c7-9f17-c3e332db911a service nova] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Refreshing network info cache for port 546b0347-51db-4a4a-99e9-4f3db96a36bd {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 919.776098] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313845, 'name': CreateVM_Task, 'duration_secs': 0.453918} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.776267] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 919.778717] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.778717] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.778717] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 919.778717] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ac06a743-f928-458f-9506-c19b685ac049 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.782752] env[62460]: DEBUG oslo_vmware.api [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Waiting for the task: (returnval){ [ 919.782752] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52194766-a608-88be-1fa4-36b7e5ed92e7" [ 919.782752] env[62460]: _type = "Task" [ 919.782752] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.791237] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquiring lock "896f8ff4-e45f-4403-a727-03ee25e58609" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.796570] env[62460]: DEBUG oslo_vmware.api [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52194766-a608-88be-1fa4-36b7e5ed92e7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.799268] env[62460]: DEBUG oslo_concurrency.lockutils [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.886557] env[62460]: DEBUG nova.network.neutron [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Successfully updated port: 3304dbf2-549a-48d4-bd32-9fbd3d98cc3d {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 919.939052] env[62460]: DEBUG oslo_concurrency.lockutils [req-0f11ea75-6eb2-433c-ba70-dcdf588014ad req-845320bf-d1b0-4c53-87e5-4b5933eb6209 service nova] Releasing lock "refresh_cache-896f8ff4-e45f-4403-a727-03ee25e58609" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.964679] env[62460]: DEBUG nova.objects.instance [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lazy-loading 'numa_topology' on Instance uuid 896f8ff4-e45f-4403-a727-03ee25e58609 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 919.985225] env[62460]: INFO nova.compute.manager [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Took 23.87 seconds to build instance. [ 919.992910] env[62460]: DEBUG oslo_concurrency.lockutils [None req-dc607155-ae9c-45e1-8539-0d765fe9cad5 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "09155629-51c8-4043-b1ee-6d5036552a67" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.851s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.296106] env[62460]: DEBUG oslo_vmware.api [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52194766-a608-88be-1fa4-36b7e5ed92e7, 'name': SearchDatastore_Task, 'duration_secs': 0.012834} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.296684] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.296936] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 920.297517] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.297709] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.297931] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 920.298787] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2ee5eba4-0671-48cb-a0b0-59e993a1ae3a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.306723] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 920.306880] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 920.307666] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dfa79957-6e4c-4d70-a66e-bbe98c9b18fd {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.313802] env[62460]: DEBUG oslo_vmware.api [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Waiting for the task: (returnval){ [ 920.313802] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]5285eaa8-1dbf-10d1-c634-0b9a64706fe4" [ 920.313802] env[62460]: _type = "Task" [ 920.313802] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.322680] env[62460]: DEBUG oslo_vmware.api [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5285eaa8-1dbf-10d1-c634-0b9a64706fe4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.391257] env[62460]: DEBUG oslo_concurrency.lockutils [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "refresh_cache-61073a4e-c5e5-48f5-a43a-ce71bd18e48a" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.391471] env[62460]: DEBUG oslo_concurrency.lockutils [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquired lock "refresh_cache-61073a4e-c5e5-48f5-a43a-ce71bd18e48a" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.391641] env[62460]: DEBUG nova.network.neutron [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 920.466228] env[62460]: DEBUG nova.objects.base [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Object Instance<896f8ff4-e45f-4403-a727-03ee25e58609> lazy-loaded attributes: resources,numa_topology {{(pid=62460) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 920.487484] env[62460]: DEBUG oslo_concurrency.lockutils [None req-02a042bd-6233-4978-ba2a-3d05e50584cd tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Lock "bf28fc98-d9a9-4678-a107-ace2b6503353" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.383s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.563500] env[62460]: DEBUG nova.network.neutron [req-ae4fcbd3-44d1-49be-ac2b-a55c0afd72a1 req-2a10b68e-9979-42c7-9f17-c3e332db911a service nova] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Updated VIF entry in instance network info cache for port 546b0347-51db-4a4a-99e9-4f3db96a36bd. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 920.564270] env[62460]: DEBUG nova.network.neutron [req-ae4fcbd3-44d1-49be-ac2b-a55c0afd72a1 req-2a10b68e-9979-42c7-9f17-c3e332db911a service nova] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Updating instance_info_cache with network_info: [{"id": "546b0347-51db-4a4a-99e9-4f3db96a36bd", "address": "fa:16:3e:54:9b:7b", "network": {"id": "d3136e32-ad55-4b73-835c-8fa5f0480767", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1511414968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53fa06008e9f43488362895e7a143700", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "054fcd1e-638e-425a-a1de-78cb188ae026", "external-id": "nsx-vlan-transportzone-658", "segmentation_id": 658, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap546b0347-51", "ovs_interfaceid": "546b0347-51db-4a4a-99e9-4f3db96a36bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 920.582870] env[62460]: DEBUG nova.compute.manager [req-aa327955-7ae0-4e20-a65e-8be0d31ec5d0 req-80b4609e-7bd3-4871-a510-5d7bced6ff01 service nova] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Received event network-vif-plugged-3304dbf2-549a-48d4-bd32-9fbd3d98cc3d {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 920.583225] env[62460]: DEBUG oslo_concurrency.lockutils [req-aa327955-7ae0-4e20-a65e-8be0d31ec5d0 req-80b4609e-7bd3-4871-a510-5d7bced6ff01 service nova] Acquiring lock "61073a4e-c5e5-48f5-a43a-ce71bd18e48a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.583502] env[62460]: DEBUG oslo_concurrency.lockutils [req-aa327955-7ae0-4e20-a65e-8be0d31ec5d0 req-80b4609e-7bd3-4871-a510-5d7bced6ff01 service nova] Lock "61073a4e-c5e5-48f5-a43a-ce71bd18e48a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.583736] env[62460]: DEBUG oslo_concurrency.lockutils [req-aa327955-7ae0-4e20-a65e-8be0d31ec5d0 req-80b4609e-7bd3-4871-a510-5d7bced6ff01 service nova] Lock "61073a4e-c5e5-48f5-a43a-ce71bd18e48a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.584018] env[62460]: DEBUG nova.compute.manager [req-aa327955-7ae0-4e20-a65e-8be0d31ec5d0 req-80b4609e-7bd3-4871-a510-5d7bced6ff01 service nova] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] No waiting events found dispatching network-vif-plugged-3304dbf2-549a-48d4-bd32-9fbd3d98cc3d {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 920.584294] env[62460]: WARNING nova.compute.manager [req-aa327955-7ae0-4e20-a65e-8be0d31ec5d0 req-80b4609e-7bd3-4871-a510-5d7bced6ff01 service nova] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Received unexpected event network-vif-plugged-3304dbf2-549a-48d4-bd32-9fbd3d98cc3d for instance with vm_state building and task_state spawning. [ 920.584533] env[62460]: DEBUG nova.compute.manager [req-aa327955-7ae0-4e20-a65e-8be0d31ec5d0 req-80b4609e-7bd3-4871-a510-5d7bced6ff01 service nova] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Received event network-changed-3304dbf2-549a-48d4-bd32-9fbd3d98cc3d {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 920.584751] env[62460]: DEBUG nova.compute.manager [req-aa327955-7ae0-4e20-a65e-8be0d31ec5d0 req-80b4609e-7bd3-4871-a510-5d7bced6ff01 service nova] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Refreshing instance network info cache due to event network-changed-3304dbf2-549a-48d4-bd32-9fbd3d98cc3d. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 920.584979] env[62460]: DEBUG oslo_concurrency.lockutils [req-aa327955-7ae0-4e20-a65e-8be0d31ec5d0 req-80b4609e-7bd3-4871-a510-5d7bced6ff01 service nova] Acquiring lock "refresh_cache-61073a4e-c5e5-48f5-a43a-ce71bd18e48a" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.645067] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-496296c6-fc88-451c-b92c-ae1d4b507aee {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.654100] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10d84d42-be10-4b27-94df-5b67789ec48c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.693283] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44043486-102f-4d62-b71f-0276f99033f6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.698400] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8088bda-04a5-4cbe-9e8c-be261344fe05 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.716231] env[62460]: DEBUG nova.compute.provider_tree [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 920.825454] env[62460]: DEBUG oslo_vmware.api [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5285eaa8-1dbf-10d1-c634-0b9a64706fe4, 'name': SearchDatastore_Task, 'duration_secs': 0.007949} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.826554] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-96facbd5-5553-4a68-a396-27af79d1ba51 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.832522] env[62460]: DEBUG oslo_vmware.api [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Waiting for the task: (returnval){ [ 920.832522] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52563101-1b59-63e8-25e6-9f535c100c07" [ 920.832522] env[62460]: _type = "Task" [ 920.832522] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.843827] env[62460]: DEBUG oslo_vmware.api [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52563101-1b59-63e8-25e6-9f535c100c07, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.930532] env[62460]: DEBUG nova.network.neutron [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 921.067356] env[62460]: DEBUG oslo_concurrency.lockutils [req-ae4fcbd3-44d1-49be-ac2b-a55c0afd72a1 req-2a10b68e-9979-42c7-9f17-c3e332db911a service nova] Releasing lock "refresh_cache-abd4501b-3ad8-4103-921d-5b80fc2f313f" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.106922] env[62460]: DEBUG nova.network.neutron [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Updating instance_info_cache with network_info: [{"id": "3304dbf2-549a-48d4-bd32-9fbd3d98cc3d", "address": "fa:16:3e:5e:9a:ad", "network": {"id": "c8b3d824-c8f0-4f7b-a91f-1987e04a5f09", "bridge": "br-int", "label": "tempest-ServersTestJSON-1769082812-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10ac54db1ef54e249a077bbb3dde4242", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "255460d5-71d4-4bfd-87f1-acc10085db7f", "external-id": "nsx-vlan-transportzone-152", "segmentation_id": 152, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3304dbf2-54", "ovs_interfaceid": "3304dbf2-549a-48d4-bd32-9fbd3d98cc3d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 921.222795] env[62460]: DEBUG nova.scheduler.client.report [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 921.345353] env[62460]: DEBUG oslo_vmware.api [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52563101-1b59-63e8-25e6-9f535c100c07, 'name': SearchDatastore_Task, 'duration_secs': 0.032322} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.345845] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.346396] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] abd4501b-3ad8-4103-921d-5b80fc2f313f/abd4501b-3ad8-4103-921d-5b80fc2f313f.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 921.352029] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-51578e54-874e-4cbb-9357-a2916250b58b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.358969] env[62460]: DEBUG oslo_vmware.api [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Waiting for the task: (returnval){ [ 921.358969] env[62460]: value = "task-1313847" [ 921.358969] env[62460]: _type = "Task" [ 921.358969] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.368196] env[62460]: DEBUG oslo_vmware.api [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313847, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.609705] env[62460]: DEBUG oslo_concurrency.lockutils [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Releasing lock "refresh_cache-61073a4e-c5e5-48f5-a43a-ce71bd18e48a" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.610049] env[62460]: DEBUG nova.compute.manager [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Instance network_info: |[{"id": "3304dbf2-549a-48d4-bd32-9fbd3d98cc3d", "address": "fa:16:3e:5e:9a:ad", "network": {"id": "c8b3d824-c8f0-4f7b-a91f-1987e04a5f09", "bridge": "br-int", "label": "tempest-ServersTestJSON-1769082812-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10ac54db1ef54e249a077bbb3dde4242", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "255460d5-71d4-4bfd-87f1-acc10085db7f", "external-id": "nsx-vlan-transportzone-152", "segmentation_id": 152, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3304dbf2-54", "ovs_interfaceid": "3304dbf2-549a-48d4-bd32-9fbd3d98cc3d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 921.610408] env[62460]: DEBUG oslo_concurrency.lockutils [req-aa327955-7ae0-4e20-a65e-8be0d31ec5d0 req-80b4609e-7bd3-4871-a510-5d7bced6ff01 service nova] Acquired lock "refresh_cache-61073a4e-c5e5-48f5-a43a-ce71bd18e48a" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.610604] env[62460]: DEBUG nova.network.neutron [req-aa327955-7ae0-4e20-a65e-8be0d31ec5d0 req-80b4609e-7bd3-4871-a510-5d7bced6ff01 service nova] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Refreshing network info cache for port 3304dbf2-549a-48d4-bd32-9fbd3d98cc3d {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 921.611899] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5e:9a:ad', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '255460d5-71d4-4bfd-87f1-acc10085db7f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3304dbf2-549a-48d4-bd32-9fbd3d98cc3d', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 921.626652] env[62460]: DEBUG oslo.service.loopingcall [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 921.631127] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 921.631789] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0c9fee72-d538-42ca-a057-18fc9262776a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.660988] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 921.660988] env[62460]: value = "task-1313848" [ 921.660988] env[62460]: _type = "Task" [ 921.660988] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.671053] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313848, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.710551] env[62460]: DEBUG nova.compute.manager [req-55d7efa4-276c-4db2-9090-dadf6b607463 req-02f1a012-91bc-44c9-815a-53a8abe54b82 service nova] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Received event network-changed-7bb00d79-430f-44e2-9e92-57109c05d0bb {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 921.710907] env[62460]: DEBUG nova.compute.manager [req-55d7efa4-276c-4db2-9090-dadf6b607463 req-02f1a012-91bc-44c9-815a-53a8abe54b82 service nova] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Refreshing instance network info cache due to event network-changed-7bb00d79-430f-44e2-9e92-57109c05d0bb. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 921.711145] env[62460]: DEBUG oslo_concurrency.lockutils [req-55d7efa4-276c-4db2-9090-dadf6b607463 req-02f1a012-91bc-44c9-815a-53a8abe54b82 service nova] Acquiring lock "refresh_cache-bf28fc98-d9a9-4678-a107-ace2b6503353" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.711952] env[62460]: DEBUG oslo_concurrency.lockutils [req-55d7efa4-276c-4db2-9090-dadf6b607463 req-02f1a012-91bc-44c9-815a-53a8abe54b82 service nova] Acquired lock "refresh_cache-bf28fc98-d9a9-4678-a107-ace2b6503353" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.711952] env[62460]: DEBUG nova.network.neutron [req-55d7efa4-276c-4db2-9090-dadf6b607463 req-02f1a012-91bc-44c9-815a-53a8abe54b82 service nova] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Refreshing network info cache for port 7bb00d79-430f-44e2-9e92-57109c05d0bb {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 921.733372] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.275s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.739054] env[62460]: DEBUG oslo_concurrency.lockutils [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 1.937s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.872921] env[62460]: DEBUG oslo_vmware.api [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313847, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.931516] env[62460]: DEBUG nova.network.neutron [req-aa327955-7ae0-4e20-a65e-8be0d31ec5d0 req-80b4609e-7bd3-4871-a510-5d7bced6ff01 service nova] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Updated VIF entry in instance network info cache for port 3304dbf2-549a-48d4-bd32-9fbd3d98cc3d. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 921.931516] env[62460]: DEBUG nova.network.neutron [req-aa327955-7ae0-4e20-a65e-8be0d31ec5d0 req-80b4609e-7bd3-4871-a510-5d7bced6ff01 service nova] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Updating instance_info_cache with network_info: [{"id": "3304dbf2-549a-48d4-bd32-9fbd3d98cc3d", "address": "fa:16:3e:5e:9a:ad", "network": {"id": "c8b3d824-c8f0-4f7b-a91f-1987e04a5f09", "bridge": "br-int", "label": "tempest-ServersTestJSON-1769082812-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10ac54db1ef54e249a077bbb3dde4242", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "255460d5-71d4-4bfd-87f1-acc10085db7f", "external-id": "nsx-vlan-transportzone-152", "segmentation_id": 152, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3304dbf2-54", "ovs_interfaceid": "3304dbf2-549a-48d4-bd32-9fbd3d98cc3d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 922.039361] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "6a5934ed-30dc-4512-8cf2-1e4d9f20836c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.039634] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "6a5934ed-30dc-4512-8cf2-1e4d9f20836c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.171699] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313848, 'name': CreateVM_Task, 'duration_secs': 0.445371} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.171699] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 922.172302] env[62460]: DEBUG oslo_concurrency.lockutils [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.172477] env[62460]: DEBUG oslo_concurrency.lockutils [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.172809] env[62460]: DEBUG oslo_concurrency.lockutils [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 922.173089] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c801f1ba-525b-4c4e-a2ab-74b652cddb5a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.177741] env[62460]: DEBUG oslo_vmware.api [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for the task: (returnval){ [ 922.177741] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52df7389-3d4c-93ed-e971-5516638014d5" [ 922.177741] env[62460]: _type = "Task" [ 922.177741] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.186611] env[62460]: DEBUG oslo_vmware.api [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52df7389-3d4c-93ed-e971-5516638014d5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.246934] env[62460]: INFO nova.compute.claims [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 922.255489] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e303f81d-5b6a-4f97-8645-e3b9ef9dd7b4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lock "896f8ff4-e45f-4403-a727-03ee25e58609" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 21.943s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.256516] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lock "896f8ff4-e45f-4403-a727-03ee25e58609" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 2.467s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.256880] env[62460]: INFO nova.compute.manager [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Unshelving [ 922.372607] env[62460]: DEBUG oslo_vmware.api [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313847, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.533857} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.372746] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] abd4501b-3ad8-4103-921d-5b80fc2f313f/abd4501b-3ad8-4103-921d-5b80fc2f313f.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 922.373203] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 922.376249] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d8688d07-ddea-44ec-9261-6bab3ad92bb0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.385136] env[62460]: DEBUG oslo_vmware.api [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Waiting for the task: (returnval){ [ 922.385136] env[62460]: value = "task-1313849" [ 922.385136] env[62460]: _type = "Task" [ 922.385136] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.393849] env[62460]: DEBUG oslo_vmware.api [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313849, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.436134] env[62460]: DEBUG oslo_concurrency.lockutils [req-aa327955-7ae0-4e20-a65e-8be0d31ec5d0 req-80b4609e-7bd3-4871-a510-5d7bced6ff01 service nova] Releasing lock "refresh_cache-61073a4e-c5e5-48f5-a43a-ce71bd18e48a" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.545563] env[62460]: DEBUG nova.compute.manager [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 922.580543] env[62460]: DEBUG nova.network.neutron [req-55d7efa4-276c-4db2-9090-dadf6b607463 req-02f1a012-91bc-44c9-815a-53a8abe54b82 service nova] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Updated VIF entry in instance network info cache for port 7bb00d79-430f-44e2-9e92-57109c05d0bb. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 922.580945] env[62460]: DEBUG nova.network.neutron [req-55d7efa4-276c-4db2-9090-dadf6b607463 req-02f1a012-91bc-44c9-815a-53a8abe54b82 service nova] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Updating instance_info_cache with network_info: [{"id": "7bb00d79-430f-44e2-9e92-57109c05d0bb", "address": "fa:16:3e:94:48:0c", "network": {"id": "e2ad0587-12a2-40ce-b2d0-c682042335fb", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1549110976-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.202", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1845e17fb7e64cf594c4e81c67ad8243", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c405e9f-a6c8-4308-acac-071654efe18e", "external-id": "nsx-vlan-transportzone-851", "segmentation_id": 851, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7bb00d79-43", "ovs_interfaceid": "7bb00d79-430f-44e2-9e92-57109c05d0bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 922.692256] env[62460]: DEBUG oslo_vmware.api [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52df7389-3d4c-93ed-e971-5516638014d5, 'name': SearchDatastore_Task, 'duration_secs': 0.009117} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.692256] env[62460]: DEBUG oslo_concurrency.lockutils [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.692256] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 922.692256] env[62460]: DEBUG oslo_concurrency.lockutils [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.692405] env[62460]: DEBUG oslo_concurrency.lockutils [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.692527] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 922.692811] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0583569a-3641-4e27-8b8e-6c0382f924f3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.702869] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 922.703083] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 922.703802] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0eb1fc35-3665-4e8d-a384-94997cb9b3c4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.709465] env[62460]: DEBUG oslo_vmware.api [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for the task: (returnval){ [ 922.709465] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]524f3b80-4e8e-ce36-20f6-e0b2f40c8629" [ 922.709465] env[62460]: _type = "Task" [ 922.709465] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.719016] env[62460]: DEBUG oslo_vmware.api [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]524f3b80-4e8e-ce36-20f6-e0b2f40c8629, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.756558] env[62460]: INFO nova.compute.resource_tracker [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Updating resource usage from migration 68cc2b47-75ac-432c-be4d-b2f8553dfe61 [ 922.779098] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd5686d8-0c05-4196-a00b-777bd0158d91 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Volume attach. Driver type: vmdk {{(pid=62460) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 922.779371] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd5686d8-0c05-4196-a00b-777bd0158d91 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281240', 'volume_id': '8131193e-aa4f-424e-b51d-e2276d84089e', 'name': 'volume-8131193e-aa4f-424e-b51d-e2276d84089e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3e10af70-db52-4d4a-bb92-821a05dcbab6', 'attached_at': '', 'detached_at': '', 'volume_id': '8131193e-aa4f-424e-b51d-e2276d84089e', 'serial': '8131193e-aa4f-424e-b51d-e2276d84089e'} {{(pid=62460) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 922.780307] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66d8a0a1-7ad3-4ef3-955e-92476aa28e2f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.801861] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0df93239-7afe-45b2-ac0f-94f550d11772 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.828216] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd5686d8-0c05-4196-a00b-777bd0158d91 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Reconfiguring VM instance instance-0000004d to attach disk [datastore2] volume-8131193e-aa4f-424e-b51d-e2276d84089e/volume-8131193e-aa4f-424e-b51d-e2276d84089e.vmdk or device None with type thin {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 922.831938] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-29b7f3de-ab69-4967-bd05-e0eef87ef1c2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.849887] env[62460]: DEBUG oslo_vmware.api [None req-bd5686d8-0c05-4196-a00b-777bd0158d91 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Waiting for the task: (returnval){ [ 922.849887] env[62460]: value = "task-1313850" [ 922.849887] env[62460]: _type = "Task" [ 922.849887] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.858921] env[62460]: DEBUG oslo_vmware.api [None req-bd5686d8-0c05-4196-a00b-777bd0158d91 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Task: {'id': task-1313850, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.896892] env[62460]: DEBUG oslo_vmware.api [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313849, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068045} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.897203] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 922.898084] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07101e17-ce68-4003-9ddf-abe0d3efa0d2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.919791] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Reconfiguring VM instance instance-00000057 to attach disk [datastore2] abd4501b-3ad8-4103-921d-5b80fc2f313f/abd4501b-3ad8-4103-921d-5b80fc2f313f.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 922.922752] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3c669b73-6c5c-4edf-a996-1f6d07a4a228 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.942858] env[62460]: DEBUG oslo_vmware.api [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Waiting for the task: (returnval){ [ 922.942858] env[62460]: value = "task-1313851" [ 922.942858] env[62460]: _type = "Task" [ 922.942858] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.953446] env[62460]: DEBUG oslo_vmware.api [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313851, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.989270] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4793c235-7477-4ea5-ad1a-9d670bd91f3d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.997474] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1c90279-f7dd-4afa-b1f1-b5867ceb62da {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.027948] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d922c8af-6193-44f6-a3f5-e8a248529a7f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.035399] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39616406-7f6a-4df2-b91d-b0f039b30005 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.050993] env[62460]: DEBUG nova.compute.provider_tree [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 923.065462] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.087174] env[62460]: DEBUG oslo_concurrency.lockutils [req-55d7efa4-276c-4db2-9090-dadf6b607463 req-02f1a012-91bc-44c9-815a-53a8abe54b82 service nova] Releasing lock "refresh_cache-bf28fc98-d9a9-4678-a107-ace2b6503353" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.222012] env[62460]: DEBUG oslo_vmware.api [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]524f3b80-4e8e-ce36-20f6-e0b2f40c8629, 'name': SearchDatastore_Task, 'duration_secs': 0.007792} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.222773] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-439065fa-5a09-4b58-ab72-ce09b11dda7e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.229430] env[62460]: DEBUG oslo_vmware.api [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for the task: (returnval){ [ 923.229430] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52b00294-15d0-4f20-813d-754872bd23d1" [ 923.229430] env[62460]: _type = "Task" [ 923.229430] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.237137] env[62460]: DEBUG oslo_vmware.api [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52b00294-15d0-4f20-813d-754872bd23d1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.241108] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Acquiring lock "9c63946c-f294-42ea-979e-68db26d88858" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.241380] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Lock "9c63946c-f294-42ea-979e-68db26d88858" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.266227] env[62460]: DEBUG nova.compute.utils [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 923.275571] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Acquiring lock "003191d3-fe38-4bde-8d7c-75b4aad413b4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.276965] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Lock "003191d3-fe38-4bde-8d7c-75b4aad413b4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.306949] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Acquiring lock "db331001-1744-4c42-a40c-87f3e60da4b8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.307218] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Lock "db331001-1744-4c42-a40c-87f3e60da4b8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.360020] env[62460]: DEBUG oslo_vmware.api [None req-bd5686d8-0c05-4196-a00b-777bd0158d91 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Task: {'id': task-1313850, 'name': ReconfigVM_Task, 'duration_secs': 0.401641} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.360843] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd5686d8-0c05-4196-a00b-777bd0158d91 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Reconfigured VM instance instance-0000004d to attach disk [datastore2] volume-8131193e-aa4f-424e-b51d-e2276d84089e/volume-8131193e-aa4f-424e-b51d-e2276d84089e.vmdk or device None with type thin {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 923.365538] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2559d256-8f41-4327-a355-5c999bb47e40 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.379930] env[62460]: DEBUG oslo_vmware.api [None req-bd5686d8-0c05-4196-a00b-777bd0158d91 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Waiting for the task: (returnval){ [ 923.379930] env[62460]: value = "task-1313852" [ 923.379930] env[62460]: _type = "Task" [ 923.379930] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.388031] env[62460]: DEBUG oslo_vmware.api [None req-bd5686d8-0c05-4196-a00b-777bd0158d91 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Task: {'id': task-1313852, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.453810] env[62460]: DEBUG oslo_vmware.api [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313851, 'name': ReconfigVM_Task, 'duration_secs': 0.348414} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.453810] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Reconfigured VM instance instance-00000057 to attach disk [datastore2] abd4501b-3ad8-4103-921d-5b80fc2f313f/abd4501b-3ad8-4103-921d-5b80fc2f313f.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 923.454471] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-97c7455f-be78-468e-b8f1-96f99eff7bea {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.460507] env[62460]: DEBUG oslo_vmware.api [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Waiting for the task: (returnval){ [ 923.460507] env[62460]: value = "task-1313853" [ 923.460507] env[62460]: _type = "Task" [ 923.460507] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.470302] env[62460]: DEBUG oslo_vmware.api [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313853, 'name': Rename_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.574088] env[62460]: ERROR nova.scheduler.client.report [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [req-486dd03f-6e9a-4a04-be68-091ae19404a8] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 76e8f00d-65cc-4766-bb9d-5006fb51c728. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-486dd03f-6e9a-4a04-be68-091ae19404a8"}]} [ 923.591744] env[62460]: DEBUG nova.scheduler.client.report [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Refreshing inventories for resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 923.606050] env[62460]: DEBUG nova.scheduler.client.report [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Updating ProviderTree inventory for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 923.606351] env[62460]: DEBUG nova.compute.provider_tree [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 923.617443] env[62460]: DEBUG nova.scheduler.client.report [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Refreshing aggregate associations for resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728, aggregates: None {{(pid=62460) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 923.634521] env[62460]: DEBUG nova.scheduler.client.report [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Refreshing trait associations for resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62460) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 923.740637] env[62460]: DEBUG oslo_vmware.api [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52b00294-15d0-4f20-813d-754872bd23d1, 'name': SearchDatastore_Task, 'duration_secs': 0.018637} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.741012] env[62460]: DEBUG oslo_concurrency.lockutils [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.741183] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] 61073a4e-c5e5-48f5-a43a-ce71bd18e48a/61073a4e-c5e5-48f5-a43a-ce71bd18e48a.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 923.741586] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ee0604da-f25f-4c67-ab0d-596dd4daf452 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.747286] env[62460]: DEBUG nova.compute.manager [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 923.751066] env[62460]: DEBUG oslo_vmware.api [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for the task: (returnval){ [ 923.751066] env[62460]: value = "task-1313854" [ 923.751066] env[62460]: _type = "Task" [ 923.751066] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.758317] env[62460]: DEBUG oslo_vmware.api [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313854, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.770932] env[62460]: INFO nova.virt.block_device [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Booting with volume c472b891-df79-428d-a905-5349e414e7f5 at /dev/sdb [ 923.778743] env[62460]: DEBUG nova.compute.manager [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 923.793723] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d4d675e-901f-4ba1-b208-b79b0742d792 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.801037] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88db9a44-e5e0-4466-baac-2d43a973f3dc {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.805967] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9b46c0d8-bc52-4c38-871c-c4ad27777c41 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.830630] env[62460]: DEBUG nova.compute.manager [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 923.834763] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2eb55c93-710d-4af6-9eee-dbbaf67c6863 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.841301] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d92fc96-9549-435e-b3f1-84047c247b7e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.854626] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3708d36e-c136-48dc-8491-06eba43ed21a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.877361] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-74b2f284-7dd9-40c2-a157-dcd29c233edb {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.879559] env[62460]: DEBUG nova.compute.provider_tree [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 923.889041] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-245867b7-70d6-43a7-b557-837b365f7d54 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.903634] env[62460]: DEBUG oslo_vmware.api [None req-bd5686d8-0c05-4196-a00b-777bd0158d91 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Task: {'id': task-1313852, 'name': ReconfigVM_Task, 'duration_secs': 0.125974} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.904407] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd5686d8-0c05-4196-a00b-777bd0158d91 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281240', 'volume_id': '8131193e-aa4f-424e-b51d-e2276d84089e', 'name': 'volume-8131193e-aa4f-424e-b51d-e2276d84089e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3e10af70-db52-4d4a-bb92-821a05dcbab6', 'attached_at': '', 'detached_at': '', 'volume_id': '8131193e-aa4f-424e-b51d-e2276d84089e', 'serial': '8131193e-aa4f-424e-b51d-e2276d84089e'} {{(pid=62460) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 923.920818] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06060df9-f2eb-4ed7-819a-cb96a51b6ad8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.928564] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3063490-8e42-4760-9ac9-132e9b810ef5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.943234] env[62460]: DEBUG nova.virt.block_device [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Updating existing volume attachment record: 0c9eb5ce-ef9b-4a0b-9d84-041b7913b082 {{(pid=62460) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 923.972059] env[62460]: DEBUG oslo_vmware.api [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313853, 'name': Rename_Task, 'duration_secs': 0.152438} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.972424] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 923.972815] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-76a80507-44d2-46a8-a58e-237e758c99e6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.980022] env[62460]: DEBUG oslo_vmware.api [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Waiting for the task: (returnval){ [ 923.980022] env[62460]: value = "task-1313855" [ 923.980022] env[62460]: _type = "Task" [ 923.980022] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.988801] env[62460]: DEBUG oslo_vmware.api [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313855, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.266502] env[62460]: DEBUG oslo_vmware.api [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313854, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.471084} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.266823] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] 61073a4e-c5e5-48f5-a43a-ce71bd18e48a/61073a4e-c5e5-48f5-a43a-ce71bd18e48a.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 924.266978] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 924.267334] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c05d80ac-62c9-421f-9d9b-b908fea06447 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.273499] env[62460]: DEBUG oslo_vmware.api [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for the task: (returnval){ [ 924.273499] env[62460]: value = "task-1313858" [ 924.273499] env[62460]: _type = "Task" [ 924.273499] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.278225] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.285386] env[62460]: DEBUG oslo_vmware.api [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313858, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.301515] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.355200] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.417505] env[62460]: DEBUG nova.scheduler.client.report [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Updated inventory for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with generation 109 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 924.418167] env[62460]: DEBUG nova.compute.provider_tree [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Updating resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 generation from 109 to 110 during operation: update_inventory {{(pid=62460) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 924.418167] env[62460]: DEBUG nova.compute.provider_tree [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 924.491806] env[62460]: DEBUG oslo_vmware.api [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313855, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.783699] env[62460]: DEBUG oslo_vmware.api [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313858, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082461} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.784043] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 924.785500] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26a3718f-8d84-4eb1-ba4d-76483ccc9cf0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.811991] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Reconfiguring VM instance instance-00000058 to attach disk [datastore2] 61073a4e-c5e5-48f5-a43a-ce71bd18e48a/61073a4e-c5e5-48f5-a43a-ce71bd18e48a.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 924.816918] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a9757695-2f0d-4960-9964-342aad5487db {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.864067] env[62460]: DEBUG oslo_vmware.api [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for the task: (returnval){ [ 924.864067] env[62460]: value = "task-1313860" [ 924.864067] env[62460]: _type = "Task" [ 924.864067] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.880727] env[62460]: DEBUG oslo_vmware.api [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313860, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.923115] env[62460]: DEBUG oslo_concurrency.lockutils [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 3.187s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.923115] env[62460]: INFO nova.compute.manager [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Migrating [ 924.923365] env[62460]: DEBUG oslo_concurrency.lockutils [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.923470] env[62460]: DEBUG oslo_concurrency.lockutils [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquired lock "compute-rpcapi-router" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.924744] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.860s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.926332] env[62460]: INFO nova.compute.claims [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 924.929412] env[62460]: INFO nova.compute.rpcapi [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Automatically selected compute RPC version 6.3 from minimum service version 67 [ 924.929891] env[62460]: DEBUG oslo_concurrency.lockutils [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Releasing lock "compute-rpcapi-router" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.944430] env[62460]: DEBUG nova.objects.instance [None req-bd5686d8-0c05-4196-a00b-777bd0158d91 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Lazy-loading 'flavor' on Instance uuid 3e10af70-db52-4d4a-bb92-821a05dcbab6 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 924.989873] env[62460]: DEBUG oslo_vmware.api [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313855, 'name': PowerOnVM_Task, 'duration_secs': 0.583474} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.990316] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 924.990510] env[62460]: INFO nova.compute.manager [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Took 8.62 seconds to spawn the instance on the hypervisor. [ 924.990759] env[62460]: DEBUG nova.compute.manager [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 924.991593] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e34a883f-278c-44e7-a4f0-1e99c88ec705 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.374326] env[62460]: DEBUG oslo_vmware.api [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313860, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.446767] env[62460]: DEBUG oslo_concurrency.lockutils [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "refresh_cache-6b084ed4-e8f0-4063-bf2a-252740640753" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.446767] env[62460]: DEBUG oslo_concurrency.lockutils [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquired lock "refresh_cache-6b084ed4-e8f0-4063-bf2a-252740640753" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.447276] env[62460]: DEBUG nova.network.neutron [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 925.451763] env[62460]: DEBUG oslo_concurrency.lockutils [None req-bd5686d8-0c05-4196-a00b-777bd0158d91 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Lock "3e10af70-db52-4d4a-bb92-821a05dcbab6" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.317s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.511230] env[62460]: INFO nova.compute.manager [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Took 26.38 seconds to build instance. [ 925.687679] env[62460]: DEBUG oslo_concurrency.lockutils [None req-8f79c83c-ecfe-4524-bc07-c2ee5fdc15f4 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Acquiring lock "3e10af70-db52-4d4a-bb92-821a05dcbab6" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.687679] env[62460]: DEBUG oslo_concurrency.lockutils [None req-8f79c83c-ecfe-4524-bc07-c2ee5fdc15f4 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Lock "3e10af70-db52-4d4a-bb92-821a05dcbab6" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.875201] env[62460]: DEBUG oslo_vmware.api [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313860, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.985165] env[62460]: DEBUG nova.compute.manager [req-68c9497a-4b6d-4b7e-a976-9452fa96d9ab req-0669b939-4f2f-4aff-a30a-2c84e5b01643 service nova] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Received event network-changed-546b0347-51db-4a4a-99e9-4f3db96a36bd {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 925.985165] env[62460]: DEBUG nova.compute.manager [req-68c9497a-4b6d-4b7e-a976-9452fa96d9ab req-0669b939-4f2f-4aff-a30a-2c84e5b01643 service nova] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Refreshing instance network info cache due to event network-changed-546b0347-51db-4a4a-99e9-4f3db96a36bd. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 925.985165] env[62460]: DEBUG oslo_concurrency.lockutils [req-68c9497a-4b6d-4b7e-a976-9452fa96d9ab req-0669b939-4f2f-4aff-a30a-2c84e5b01643 service nova] Acquiring lock "refresh_cache-abd4501b-3ad8-4103-921d-5b80fc2f313f" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.985165] env[62460]: DEBUG oslo_concurrency.lockutils [req-68c9497a-4b6d-4b7e-a976-9452fa96d9ab req-0669b939-4f2f-4aff-a30a-2c84e5b01643 service nova] Acquired lock "refresh_cache-abd4501b-3ad8-4103-921d-5b80fc2f313f" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.985165] env[62460]: DEBUG nova.network.neutron [req-68c9497a-4b6d-4b7e-a976-9452fa96d9ab req-0669b939-4f2f-4aff-a30a-2c84e5b01643 service nova] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Refreshing network info cache for port 546b0347-51db-4a4a-99e9-4f3db96a36bd {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 926.014899] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4dcc1ca8-e54b-42d8-9e1f-37c7e4c3fd94 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "abd4501b-3ad8-4103-921d-5b80fc2f313f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.897s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.136354] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cc6da2b-d552-47bb-89ae-4e63ca83f72e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.144951] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f40a019c-4587-4458-8ba5-545737cb96c2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.178891] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c516998-d5c2-422d-abc6-a699dab4ea0d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.186575] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03e51ddd-cf53-4b71-a999-9e6624b94702 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.190933] env[62460]: INFO nova.compute.manager [None req-8f79c83c-ecfe-4524-bc07-c2ee5fdc15f4 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Detaching volume 8131193e-aa4f-424e-b51d-e2276d84089e [ 926.204246] env[62460]: DEBUG nova.compute.provider_tree [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 926.244579] env[62460]: INFO nova.virt.block_device [None req-8f79c83c-ecfe-4524-bc07-c2ee5fdc15f4 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Attempting to driver detach volume 8131193e-aa4f-424e-b51d-e2276d84089e from mountpoint /dev/sdb [ 926.244812] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f79c83c-ecfe-4524-bc07-c2ee5fdc15f4 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Volume detach. Driver type: vmdk {{(pid=62460) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 926.245037] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f79c83c-ecfe-4524-bc07-c2ee5fdc15f4 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281240', 'volume_id': '8131193e-aa4f-424e-b51d-e2276d84089e', 'name': 'volume-8131193e-aa4f-424e-b51d-e2276d84089e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3e10af70-db52-4d4a-bb92-821a05dcbab6', 'attached_at': '', 'detached_at': '', 'volume_id': '8131193e-aa4f-424e-b51d-e2276d84089e', 'serial': '8131193e-aa4f-424e-b51d-e2276d84089e'} {{(pid=62460) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 926.245959] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e69a056-4eab-435e-b43a-0cd1e6e16a85 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.269315] env[62460]: DEBUG nova.network.neutron [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Updating instance_info_cache with network_info: [{"id": "3eaf04a4-feda-4a00-908e-d121fc775ecb", "address": "fa:16:3e:98:27:28", "network": {"id": "3c49720c-94c9-4e2f-bbb7-4266cc97faee", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1650281790-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfa5a65b0f614e769de5b3aa77bf869e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3eaf04a4-fe", "ovs_interfaceid": "3eaf04a4-feda-4a00-908e-d121fc775ecb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.271038] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfe5de0e-51f8-4d89-9887-bf959f983650 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.278837] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebd5bd31-9293-4c80-9ad9-c9163c8b4427 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.300157] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37e30579-1561-48c8-861d-1a282f0d11f1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.316452] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f79c83c-ecfe-4524-bc07-c2ee5fdc15f4 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] The volume has not been displaced from its original location: [datastore2] volume-8131193e-aa4f-424e-b51d-e2276d84089e/volume-8131193e-aa4f-424e-b51d-e2276d84089e.vmdk. No consolidation needed. {{(pid=62460) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 926.322693] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f79c83c-ecfe-4524-bc07-c2ee5fdc15f4 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Reconfiguring VM instance instance-0000004d to detach disk 2001 {{(pid=62460) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 926.323032] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a81a9b50-50bc-4a02-a0ca-53f428b5be42 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.343519] env[62460]: DEBUG oslo_vmware.api [None req-8f79c83c-ecfe-4524-bc07-c2ee5fdc15f4 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Waiting for the task: (returnval){ [ 926.343519] env[62460]: value = "task-1313861" [ 926.343519] env[62460]: _type = "Task" [ 926.343519] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.351986] env[62460]: DEBUG oslo_vmware.api [None req-8f79c83c-ecfe-4524-bc07-c2ee5fdc15f4 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Task: {'id': task-1313861, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.376317] env[62460]: DEBUG oslo_vmware.api [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313860, 'name': ReconfigVM_Task, 'duration_secs': 1.104195} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.376663] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Reconfigured VM instance instance-00000058 to attach disk [datastore2] 61073a4e-c5e5-48f5-a43a-ce71bd18e48a/61073a4e-c5e5-48f5-a43a-ce71bd18e48a.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 926.377439] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-510bd1d7-5b5e-4b13-b2a4-d6d90811ff84 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.385742] env[62460]: DEBUG oslo_vmware.api [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for the task: (returnval){ [ 926.385742] env[62460]: value = "task-1313863" [ 926.385742] env[62460]: _type = "Task" [ 926.385742] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.395364] env[62460]: DEBUG oslo_vmware.api [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313863, 'name': Rename_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.709452] env[62460]: DEBUG nova.scheduler.client.report [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 926.716404] env[62460]: DEBUG nova.network.neutron [req-68c9497a-4b6d-4b7e-a976-9452fa96d9ab req-0669b939-4f2f-4aff-a30a-2c84e5b01643 service nova] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Updated VIF entry in instance network info cache for port 546b0347-51db-4a4a-99e9-4f3db96a36bd. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 926.716744] env[62460]: DEBUG nova.network.neutron [req-68c9497a-4b6d-4b7e-a976-9452fa96d9ab req-0669b939-4f2f-4aff-a30a-2c84e5b01643 service nova] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Updating instance_info_cache with network_info: [{"id": "546b0347-51db-4a4a-99e9-4f3db96a36bd", "address": "fa:16:3e:54:9b:7b", "network": {"id": "d3136e32-ad55-4b73-835c-8fa5f0480767", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1511414968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53fa06008e9f43488362895e7a143700", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "054fcd1e-638e-425a-a1de-78cb188ae026", "external-id": "nsx-vlan-transportzone-658", "segmentation_id": 658, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap546b0347-51", "ovs_interfaceid": "546b0347-51db-4a4a-99e9-4f3db96a36bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.774691] env[62460]: DEBUG oslo_concurrency.lockutils [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Releasing lock "refresh_cache-6b084ed4-e8f0-4063-bf2a-252740640753" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.855307] env[62460]: DEBUG oslo_vmware.api [None req-8f79c83c-ecfe-4524-bc07-c2ee5fdc15f4 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Task: {'id': task-1313861, 'name': ReconfigVM_Task, 'duration_secs': 0.234518} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.855598] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f79c83c-ecfe-4524-bc07-c2ee5fdc15f4 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Reconfigured VM instance instance-0000004d to detach disk 2001 {{(pid=62460) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 926.860170] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2d58e6eb-fd20-4cd0-b001-3a6277cf5072 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.874200] env[62460]: DEBUG oslo_vmware.api [None req-8f79c83c-ecfe-4524-bc07-c2ee5fdc15f4 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Waiting for the task: (returnval){ [ 926.874200] env[62460]: value = "task-1313864" [ 926.874200] env[62460]: _type = "Task" [ 926.874200] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.881694] env[62460]: DEBUG oslo_concurrency.lockutils [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.882243] env[62460]: DEBUG oslo_concurrency.lockutils [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.886832] env[62460]: DEBUG oslo_vmware.api [None req-8f79c83c-ecfe-4524-bc07-c2ee5fdc15f4 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Task: {'id': task-1313864, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.897352] env[62460]: DEBUG oslo_vmware.api [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313863, 'name': Rename_Task, 'duration_secs': 0.139941} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.897632] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 926.897880] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-718230cb-1884-482d-ab25-2a19163779ea {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.904697] env[62460]: DEBUG oslo_vmware.api [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for the task: (returnval){ [ 926.904697] env[62460]: value = "task-1313865" [ 926.904697] env[62460]: _type = "Task" [ 926.904697] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.912438] env[62460]: DEBUG oslo_vmware.api [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313865, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.215534] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.290s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.215758] env[62460]: DEBUG nova.compute.manager [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 927.218742] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.941s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.220660] env[62460]: INFO nova.compute.claims [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 927.223483] env[62460]: DEBUG oslo_concurrency.lockutils [req-68c9497a-4b6d-4b7e-a976-9452fa96d9ab req-0669b939-4f2f-4aff-a30a-2c84e5b01643 service nova] Releasing lock "refresh_cache-abd4501b-3ad8-4103-921d-5b80fc2f313f" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.384274] env[62460]: DEBUG oslo_vmware.api [None req-8f79c83c-ecfe-4524-bc07-c2ee5fdc15f4 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Task: {'id': task-1313864, 'name': ReconfigVM_Task, 'duration_secs': 0.155519} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.384620] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f79c83c-ecfe-4524-bc07-c2ee5fdc15f4 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281240', 'volume_id': '8131193e-aa4f-424e-b51d-e2276d84089e', 'name': 'volume-8131193e-aa4f-424e-b51d-e2276d84089e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3e10af70-db52-4d4a-bb92-821a05dcbab6', 'attached_at': '', 'detached_at': '', 'volume_id': '8131193e-aa4f-424e-b51d-e2276d84089e', 'serial': '8131193e-aa4f-424e-b51d-e2276d84089e'} {{(pid=62460) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 927.388327] env[62460]: DEBUG nova.compute.manager [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 927.415561] env[62460]: DEBUG oslo_vmware.api [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313865, 'name': PowerOnVM_Task, 'duration_secs': 0.504579} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.415838] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 927.417078] env[62460]: INFO nova.compute.manager [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Took 8.76 seconds to spawn the instance on the hypervisor. [ 927.417078] env[62460]: DEBUG nova.compute.manager [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 927.417288] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46f4b817-e1fb-49a3-8982-6dd96b6f094e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.728888] env[62460]: DEBUG nova.compute.utils [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 927.729911] env[62460]: DEBUG nova.compute.manager [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 927.730112] env[62460]: DEBUG nova.network.neutron [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 927.760868] env[62460]: DEBUG oslo_concurrency.lockutils [None req-25c7b810-0a7f-4b89-8417-4bb333e05885 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Acquiring lock "5b6e8205-003e-49c0-a73d-be2e032a8272" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.761187] env[62460]: DEBUG oslo_concurrency.lockutils [None req-25c7b810-0a7f-4b89-8417-4bb333e05885 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lock "5b6e8205-003e-49c0-a73d-be2e032a8272" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.776333] env[62460]: DEBUG nova.policy [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5f69b3e3f50248ff8e57054294a04686', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0da056d93bdf40c39d6e82e457727ff6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 927.910988] env[62460]: DEBUG oslo_concurrency.lockutils [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.935486] env[62460]: DEBUG nova.objects.instance [None req-8f79c83c-ecfe-4524-bc07-c2ee5fdc15f4 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Lazy-loading 'flavor' on Instance uuid 3e10af70-db52-4d4a-bb92-821a05dcbab6 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 927.938069] env[62460]: INFO nova.compute.manager [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Took 21.73 seconds to build instance. [ 928.089832] env[62460]: DEBUG nova.network.neutron [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] Successfully created port: 555d2fbe-16ac-4dd3-baa7-5b3b04b75670 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 928.234834] env[62460]: DEBUG nova.compute.manager [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 928.266627] env[62460]: INFO nova.compute.manager [None req-25c7b810-0a7f-4b89-8417-4bb333e05885 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Detaching volume a05c7ae1-7cd9-447e-8453-15a2bcde137e [ 928.291591] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-536edf6c-7a6c-42dc-91f2-ed39e259c4a9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.312859] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Updating instance '6b084ed4-e8f0-4063-bf2a-252740640753' progress to 0 {{(pid=62460) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 928.317228] env[62460]: INFO nova.virt.block_device [None req-25c7b810-0a7f-4b89-8417-4bb333e05885 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Attempting to driver detach volume a05c7ae1-7cd9-447e-8453-15a2bcde137e from mountpoint /dev/sdb [ 928.317451] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-25c7b810-0a7f-4b89-8417-4bb333e05885 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Volume detach. Driver type: vmdk {{(pid=62460) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 928.317648] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-25c7b810-0a7f-4b89-8417-4bb333e05885 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281224', 'volume_id': 'a05c7ae1-7cd9-447e-8453-15a2bcde137e', 'name': 'volume-a05c7ae1-7cd9-447e-8453-15a2bcde137e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5b6e8205-003e-49c0-a73d-be2e032a8272', 'attached_at': '', 'detached_at': '', 'volume_id': 'a05c7ae1-7cd9-447e-8453-15a2bcde137e', 'serial': 'a05c7ae1-7cd9-447e-8453-15a2bcde137e'} {{(pid=62460) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 928.324020] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79cae5c9-0835-4438-a17b-b654bd3f636e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.347680] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-984926d4-9aac-40fe-9bce-46661b44c03c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.355420] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b2a066f-cf40-47a0-ba7c-2357e8467c9c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.381936] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd8fe67a-4618-4139-b676-dc8abd54ba9a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.406953] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-25c7b810-0a7f-4b89-8417-4bb333e05885 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] The volume has not been displaced from its original location: [datastore1] volume-a05c7ae1-7cd9-447e-8453-15a2bcde137e/volume-a05c7ae1-7cd9-447e-8453-15a2bcde137e.vmdk. No consolidation needed. {{(pid=62460) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 928.412553] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-25c7b810-0a7f-4b89-8417-4bb333e05885 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Reconfiguring VM instance instance-00000042 to detach disk 2001 {{(pid=62460) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 928.415500] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8a5e68fd-7487-4909-937a-0a0345bdd00f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.435056] env[62460]: DEBUG oslo_vmware.api [None req-25c7b810-0a7f-4b89-8417-4bb333e05885 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Waiting for the task: (returnval){ [ 928.435056] env[62460]: value = "task-1313866" [ 928.435056] env[62460]: _type = "Task" [ 928.435056] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.443547] env[62460]: DEBUG oslo_concurrency.lockutils [None req-18e7cb05-8d2f-4291-b71e-38554c68fb5a tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "61073a4e-c5e5-48f5-a43a-ce71bd18e48a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.243s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.450650] env[62460]: DEBUG oslo_vmware.api [None req-25c7b810-0a7f-4b89-8417-4bb333e05885 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1313866, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.516014] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-015cc575-dcdc-45c2-9ec1-6677b76c3558 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.523598] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e844a06-65c6-4678-b9f0-6290550b8d0a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.553689] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38f42ce7-258f-4313-b765-55ecad7fd366 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.562334] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54a46d02-5c33-4ede-a0cd-9ec8673b6a93 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.579783] env[62460]: DEBUG nova.compute.provider_tree [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 928.827054] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 928.827229] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-68a0c8a1-7508-4288-9b14-4fe1d7c2f97b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.837950] env[62460]: DEBUG oslo_vmware.api [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 928.837950] env[62460]: value = "task-1313867" [ 928.837950] env[62460]: _type = "Task" [ 928.837950] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.846378] env[62460]: DEBUG oslo_vmware.api [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313867, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.946059] env[62460]: DEBUG oslo_vmware.api [None req-25c7b810-0a7f-4b89-8417-4bb333e05885 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1313866, 'name': ReconfigVM_Task, 'duration_secs': 0.44081} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.946412] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-25c7b810-0a7f-4b89-8417-4bb333e05885 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Reconfigured VM instance instance-00000042 to detach disk 2001 {{(pid=62460) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 928.951684] env[62460]: DEBUG oslo_concurrency.lockutils [None req-8f79c83c-ecfe-4524-bc07-c2ee5fdc15f4 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Lock "3e10af70-db52-4d4a-bb92-821a05dcbab6" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.265s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.952780] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-994d7955-1f01-4438-a21b-30febee53003 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.968659] env[62460]: DEBUG oslo_vmware.api [None req-25c7b810-0a7f-4b89-8417-4bb333e05885 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Waiting for the task: (returnval){ [ 928.968659] env[62460]: value = "task-1313868" [ 928.968659] env[62460]: _type = "Task" [ 928.968659] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.977437] env[62460]: DEBUG oslo_vmware.api [None req-25c7b810-0a7f-4b89-8417-4bb333e05885 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1313868, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.082754] env[62460]: DEBUG oslo_concurrency.lockutils [None req-800eb51c-26fc-48d3-9ea6-3689387e9e69 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "61073a4e-c5e5-48f5-a43a-ce71bd18e48a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.082941] env[62460]: DEBUG oslo_concurrency.lockutils [None req-800eb51c-26fc-48d3-9ea6-3689387e9e69 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "61073a4e-c5e5-48f5-a43a-ce71bd18e48a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.083225] env[62460]: DEBUG oslo_concurrency.lockutils [None req-800eb51c-26fc-48d3-9ea6-3689387e9e69 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "61073a4e-c5e5-48f5-a43a-ce71bd18e48a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.083385] env[62460]: DEBUG oslo_concurrency.lockutils [None req-800eb51c-26fc-48d3-9ea6-3689387e9e69 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "61073a4e-c5e5-48f5-a43a-ce71bd18e48a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.083593] env[62460]: DEBUG oslo_concurrency.lockutils [None req-800eb51c-26fc-48d3-9ea6-3689387e9e69 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "61073a4e-c5e5-48f5-a43a-ce71bd18e48a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.085840] env[62460]: DEBUG nova.scheduler.client.report [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 929.089401] env[62460]: INFO nova.compute.manager [None req-800eb51c-26fc-48d3-9ea6-3689387e9e69 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Terminating instance [ 929.091705] env[62460]: DEBUG nova.compute.manager [None req-800eb51c-26fc-48d3-9ea6-3689387e9e69 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 929.091973] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-800eb51c-26fc-48d3-9ea6-3689387e9e69 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 929.093113] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90b582ba-b4e7-4ea7-9ecf-276a1d15a4d3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.103814] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-800eb51c-26fc-48d3-9ea6-3689387e9e69 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 929.104207] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0a5e3c4c-2688-4962-ba8f-eab27e72b88c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.114236] env[62460]: DEBUG oslo_vmware.api [None req-800eb51c-26fc-48d3-9ea6-3689387e9e69 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for the task: (returnval){ [ 929.114236] env[62460]: value = "task-1313869" [ 929.114236] env[62460]: _type = "Task" [ 929.114236] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.122969] env[62460]: DEBUG oslo_vmware.api [None req-800eb51c-26fc-48d3-9ea6-3689387e9e69 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313869, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.243854] env[62460]: DEBUG nova.compute.manager [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 929.269335] env[62460]: DEBUG nova.virt.hardware [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 929.269605] env[62460]: DEBUG nova.virt.hardware [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 929.269771] env[62460]: DEBUG nova.virt.hardware [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 929.269961] env[62460]: DEBUG nova.virt.hardware [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 929.270138] env[62460]: DEBUG nova.virt.hardware [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 929.270313] env[62460]: DEBUG nova.virt.hardware [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 929.270534] env[62460]: DEBUG nova.virt.hardware [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 929.270702] env[62460]: DEBUG nova.virt.hardware [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 929.270878] env[62460]: DEBUG nova.virt.hardware [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 929.271061] env[62460]: DEBUG nova.virt.hardware [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 929.271249] env[62460]: DEBUG nova.virt.hardware [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 929.272097] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a35fafc6-65a8-4d8e-ba57-b3df093ff5de {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.279640] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cc8ded7-a6c4-40b4-8e72-3dfe1de16d40 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.347637] env[62460]: DEBUG oslo_vmware.api [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313867, 'name': PowerOffVM_Task, 'duration_secs': 0.316854} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.347982] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 929.348261] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Updating instance '6b084ed4-e8f0-4063-bf2a-252740640753' progress to 17 {{(pid=62460) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 929.478807] env[62460]: DEBUG oslo_vmware.api [None req-25c7b810-0a7f-4b89-8417-4bb333e05885 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1313868, 'name': ReconfigVM_Task, 'duration_secs': 0.151806} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.479081] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-25c7b810-0a7f-4b89-8417-4bb333e05885 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281224', 'volume_id': 'a05c7ae1-7cd9-447e-8453-15a2bcde137e', 'name': 'volume-a05c7ae1-7cd9-447e-8453-15a2bcde137e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5b6e8205-003e-49c0-a73d-be2e032a8272', 'attached_at': '', 'detached_at': '', 'volume_id': 'a05c7ae1-7cd9-447e-8453-15a2bcde137e', 'serial': 'a05c7ae1-7cd9-447e-8453-15a2bcde137e'} {{(pid=62460) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 929.548806] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.577273] env[62460]: DEBUG nova.compute.manager [req-2c3083b6-0a89-4bc7-9ebb-4d17d2b3a66e req-ca14a34b-0207-4b18-9865-04bb190b93ba service nova] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] Received event network-vif-plugged-555d2fbe-16ac-4dd3-baa7-5b3b04b75670 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 929.577273] env[62460]: DEBUG oslo_concurrency.lockutils [req-2c3083b6-0a89-4bc7-9ebb-4d17d2b3a66e req-ca14a34b-0207-4b18-9865-04bb190b93ba service nova] Acquiring lock "6a5934ed-30dc-4512-8cf2-1e4d9f20836c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.579523] env[62460]: DEBUG oslo_concurrency.lockutils [req-2c3083b6-0a89-4bc7-9ebb-4d17d2b3a66e req-ca14a34b-0207-4b18-9865-04bb190b93ba service nova] Lock "6a5934ed-30dc-4512-8cf2-1e4d9f20836c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.003s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.579884] env[62460]: DEBUG oslo_concurrency.lockutils [req-2c3083b6-0a89-4bc7-9ebb-4d17d2b3a66e req-ca14a34b-0207-4b18-9865-04bb190b93ba service nova] Lock "6a5934ed-30dc-4512-8cf2-1e4d9f20836c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.580221] env[62460]: DEBUG nova.compute.manager [req-2c3083b6-0a89-4bc7-9ebb-4d17d2b3a66e req-ca14a34b-0207-4b18-9865-04bb190b93ba service nova] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] No waiting events found dispatching network-vif-plugged-555d2fbe-16ac-4dd3-baa7-5b3b04b75670 {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 929.581085] env[62460]: WARNING nova.compute.manager [req-2c3083b6-0a89-4bc7-9ebb-4d17d2b3a66e req-ca14a34b-0207-4b18-9865-04bb190b93ba service nova] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] Received unexpected event network-vif-plugged-555d2fbe-16ac-4dd3-baa7-5b3b04b75670 for instance with vm_state building and task_state spawning. [ 929.595100] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.374s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.595100] env[62460]: DEBUG nova.compute.manager [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 929.596161] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.295s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.597837] env[62460]: INFO nova.compute.claims [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 929.626386] env[62460]: DEBUG oslo_vmware.api [None req-800eb51c-26fc-48d3-9ea6-3689387e9e69 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313869, 'name': PowerOffVM_Task, 'duration_secs': 0.341115} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.627674] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-800eb51c-26fc-48d3-9ea6-3689387e9e69 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 929.627674] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-800eb51c-26fc-48d3-9ea6-3689387e9e69 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 929.628299] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9a28f364-652d-4fde-83bd-75a12206fd75 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.657049] env[62460]: DEBUG nova.network.neutron [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] Successfully updated port: 555d2fbe-16ac-4dd3-baa7-5b3b04b75670 {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 929.705145] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-800eb51c-26fc-48d3-9ea6-3689387e9e69 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 929.705145] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-800eb51c-26fc-48d3-9ea6-3689387e9e69 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Deleting contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 929.705145] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-800eb51c-26fc-48d3-9ea6-3689387e9e69 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Deleting the datastore file [datastore2] 61073a4e-c5e5-48f5-a43a-ce71bd18e48a {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 929.705438] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-77fccda9-fb36-4cc2-b3a2-55e7f5852bd4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.714862] env[62460]: DEBUG oslo_vmware.api [None req-800eb51c-26fc-48d3-9ea6-3689387e9e69 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for the task: (returnval){ [ 929.714862] env[62460]: value = "task-1313871" [ 929.714862] env[62460]: _type = "Task" [ 929.714862] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.722966] env[62460]: DEBUG oslo_vmware.api [None req-800eb51c-26fc-48d3-9ea6-3689387e9e69 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313871, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.857581] env[62460]: DEBUG nova.virt.hardware [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 929.858194] env[62460]: DEBUG nova.virt.hardware [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 929.858259] env[62460]: DEBUG nova.virt.hardware [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 929.858442] env[62460]: DEBUG nova.virt.hardware [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 929.858608] env[62460]: DEBUG nova.virt.hardware [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 929.858766] env[62460]: DEBUG nova.virt.hardware [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 929.858982] env[62460]: DEBUG nova.virt.hardware [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 929.859241] env[62460]: DEBUG nova.virt.hardware [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 929.859496] env[62460]: DEBUG nova.virt.hardware [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 929.859739] env[62460]: DEBUG nova.virt.hardware [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 929.859935] env[62460]: DEBUG nova.virt.hardware [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 929.864997] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a1b95f72-7734-4481-b8c1-3b96e1149512 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.881655] env[62460]: DEBUG oslo_vmware.api [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 929.881655] env[62460]: value = "task-1313872" [ 929.881655] env[62460]: _type = "Task" [ 929.881655] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.889884] env[62460]: DEBUG oslo_vmware.api [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313872, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.032997] env[62460]: DEBUG nova.objects.instance [None req-25c7b810-0a7f-4b89-8417-4bb333e05885 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lazy-loading 'flavor' on Instance uuid 5b6e8205-003e-49c0-a73d-be2e032a8272 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 930.035221] env[62460]: DEBUG oslo_concurrency.lockutils [None req-18074529-fe3b-4b5f-9546-b5cf5e11a600 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Acquiring lock "3e10af70-db52-4d4a-bb92-821a05dcbab6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.035484] env[62460]: DEBUG oslo_concurrency.lockutils [None req-18074529-fe3b-4b5f-9546-b5cf5e11a600 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Lock "3e10af70-db52-4d4a-bb92-821a05dcbab6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.035697] env[62460]: DEBUG oslo_concurrency.lockutils [None req-18074529-fe3b-4b5f-9546-b5cf5e11a600 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Acquiring lock "3e10af70-db52-4d4a-bb92-821a05dcbab6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.035885] env[62460]: DEBUG oslo_concurrency.lockutils [None req-18074529-fe3b-4b5f-9546-b5cf5e11a600 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Lock "3e10af70-db52-4d4a-bb92-821a05dcbab6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.036075] env[62460]: DEBUG oslo_concurrency.lockutils [None req-18074529-fe3b-4b5f-9546-b5cf5e11a600 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Lock "3e10af70-db52-4d4a-bb92-821a05dcbab6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.037885] env[62460]: INFO nova.compute.manager [None req-18074529-fe3b-4b5f-9546-b5cf5e11a600 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Terminating instance [ 930.039547] env[62460]: DEBUG nova.compute.manager [None req-18074529-fe3b-4b5f-9546-b5cf5e11a600 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 930.039756] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-18074529-fe3b-4b5f-9546-b5cf5e11a600 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 930.040681] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99b6ddc3-299d-4690-9898-bc7579b6feed {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.048475] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-18074529-fe3b-4b5f-9546-b5cf5e11a600 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 930.049500] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b2ee9bfe-b741-4b83-b9b7-0409e00a933b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.054997] env[62460]: DEBUG oslo_vmware.api [None req-18074529-fe3b-4b5f-9546-b5cf5e11a600 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Waiting for the task: (returnval){ [ 930.054997] env[62460]: value = "task-1313873" [ 930.054997] env[62460]: _type = "Task" [ 930.054997] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.063394] env[62460]: DEBUG oslo_vmware.api [None req-18074529-fe3b-4b5f-9546-b5cf5e11a600 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Task: {'id': task-1313873, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.098090] env[62460]: DEBUG nova.compute.utils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 930.100286] env[62460]: DEBUG nova.compute.manager [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 930.100528] env[62460]: DEBUG nova.network.neutron [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 9c63946c-f294-42ea-979e-68db26d88858] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 930.161433] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "refresh_cache-6a5934ed-30dc-4512-8cf2-1e4d9f20836c" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.161626] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquired lock "refresh_cache-6a5934ed-30dc-4512-8cf2-1e4d9f20836c" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.161854] env[62460]: DEBUG nova.network.neutron [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 930.175195] env[62460]: DEBUG nova.policy [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '13ee01926a10432dbb135f08d7924da3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '28e73bd27f474a81a0aa7d4bb2d8efb4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 930.225166] env[62460]: DEBUG oslo_vmware.api [None req-800eb51c-26fc-48d3-9ea6-3689387e9e69 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313871, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.217806} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.225439] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-800eb51c-26fc-48d3-9ea6-3689387e9e69 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 930.225633] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-800eb51c-26fc-48d3-9ea6-3689387e9e69 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Deleted contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 930.225835] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-800eb51c-26fc-48d3-9ea6-3689387e9e69 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 930.226073] env[62460]: INFO nova.compute.manager [None req-800eb51c-26fc-48d3-9ea6-3689387e9e69 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Took 1.13 seconds to destroy the instance on the hypervisor. [ 930.226368] env[62460]: DEBUG oslo.service.loopingcall [None req-800eb51c-26fc-48d3-9ea6-3689387e9e69 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 930.226578] env[62460]: DEBUG nova.compute.manager [-] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 930.226686] env[62460]: DEBUG nova.network.neutron [-] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 930.399091] env[62460]: DEBUG oslo_vmware.api [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313872, 'name': ReconfigVM_Task, 'duration_secs': 0.175925} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.402544] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Updating instance '6b084ed4-e8f0-4063-bf2a-252740640753' progress to 33 {{(pid=62460) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 930.529326] env[62460]: DEBUG nova.compute.manager [req-5d427f79-ca5c-46a9-9ceb-9387bdae23a5 req-0ed3cd3d-1c13-4886-8ab5-40c61b8a1341 service nova] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Received event network-vif-deleted-3304dbf2-549a-48d4-bd32-9fbd3d98cc3d {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 930.529599] env[62460]: INFO nova.compute.manager [req-5d427f79-ca5c-46a9-9ceb-9387bdae23a5 req-0ed3cd3d-1c13-4886-8ab5-40c61b8a1341 service nova] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Neutron deleted interface 3304dbf2-549a-48d4-bd32-9fbd3d98cc3d; detaching it from the instance and deleting it from the info cache [ 930.529823] env[62460]: DEBUG nova.network.neutron [req-5d427f79-ca5c-46a9-9ceb-9387bdae23a5 req-0ed3cd3d-1c13-4886-8ab5-40c61b8a1341 service nova] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 930.564780] env[62460]: DEBUG oslo_vmware.api [None req-18074529-fe3b-4b5f-9546-b5cf5e11a600 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Task: {'id': task-1313873, 'name': PowerOffVM_Task, 'duration_secs': 0.187509} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.564997] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-18074529-fe3b-4b5f-9546-b5cf5e11a600 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 930.566561] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-18074529-fe3b-4b5f-9546-b5cf5e11a600 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 930.566729] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fcf11f8c-2b30-4742-8f90-63fe43c6958e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.600862] env[62460]: DEBUG nova.compute.manager [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 930.624833] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-18074529-fe3b-4b5f-9546-b5cf5e11a600 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 930.625096] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-18074529-fe3b-4b5f-9546-b5cf5e11a600 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Deleting contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 930.625300] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-18074529-fe3b-4b5f-9546-b5cf5e11a600 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Deleting the datastore file [datastore1] 3e10af70-db52-4d4a-bb92-821a05dcbab6 {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 930.625565] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-98868f90-e9bf-4aef-8947-b9096170d6c1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.639343] env[62460]: DEBUG oslo_vmware.api [None req-18074529-fe3b-4b5f-9546-b5cf5e11a600 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Waiting for the task: (returnval){ [ 930.639343] env[62460]: value = "task-1313875" [ 930.639343] env[62460]: _type = "Task" [ 930.639343] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.649103] env[62460]: DEBUG oslo_vmware.api [None req-18074529-fe3b-4b5f-9546-b5cf5e11a600 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Task: {'id': task-1313875, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.705785] env[62460]: DEBUG nova.network.neutron [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 930.718139] env[62460]: DEBUG nova.network.neutron [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Successfully created port: 2770c44d-3afd-4465-84ff-de06bb9b1a18 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 930.838624] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dec90551-2c9d-439d-a57f-e0d2f548f6fb {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.846396] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87b63fce-04dd-4ae5-b7df-cfa02d4ef4f5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.875901] env[62460]: DEBUG nova.network.neutron [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] Updating instance_info_cache with network_info: [{"id": "555d2fbe-16ac-4dd3-baa7-5b3b04b75670", "address": "fa:16:3e:f1:f5:35", "network": {"id": "de789887-9cf9-4614-a7e1-dab430aaabaf", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-656400207-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0da056d93bdf40c39d6e82e457727ff6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5116f690-f825-4fee-8a47-42b073e716c5", "external-id": "nsx-vlan-transportzone-692", "segmentation_id": 692, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap555d2fbe-16", "ovs_interfaceid": "555d2fbe-16ac-4dd3-baa7-5b3b04b75670", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 930.878188] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-442a92d2-1b68-4ee5-a397-0ce54e1ce953 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.888392] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0dbb3b1-b9eb-4e26-b087-0aff7e409d70 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.903031] env[62460]: DEBUG nova.compute.provider_tree [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 930.911154] env[62460]: DEBUG nova.virt.hardware [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 930.911382] env[62460]: DEBUG nova.virt.hardware [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 930.911550] env[62460]: DEBUG nova.virt.hardware [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 930.911736] env[62460]: DEBUG nova.virt.hardware [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 930.911888] env[62460]: DEBUG nova.virt.hardware [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 930.912051] env[62460]: DEBUG nova.virt.hardware [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 930.912264] env[62460]: DEBUG nova.virt.hardware [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 930.912439] env[62460]: DEBUG nova.virt.hardware [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 930.912608] env[62460]: DEBUG nova.virt.hardware [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 930.912774] env[62460]: DEBUG nova.virt.hardware [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 930.912952] env[62460]: DEBUG nova.virt.hardware [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 930.918515] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Reconfiguring VM instance instance-00000055 to detach disk 2000 {{(pid=62460) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 930.919385] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-be13dc86-96eb-4f00-8b04-a2f0b31579fa {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.936816] env[62460]: DEBUG oslo_vmware.api [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 930.936816] env[62460]: value = "task-1313876" [ 930.936816] env[62460]: _type = "Task" [ 930.936816] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.945276] env[62460]: DEBUG oslo_vmware.api [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313876, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.985698] env[62460]: DEBUG oslo_concurrency.lockutils [None req-87ee004f-61ca-441d-9f0d-e80687167704 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Acquiring lock "5b6e8205-003e-49c0-a73d-be2e032a8272" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.003781] env[62460]: DEBUG nova.network.neutron [-] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.034012] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d9def6f3-eddb-4a8d-a967-f853d24bb651 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.040518] env[62460]: DEBUG oslo_concurrency.lockutils [None req-25c7b810-0a7f-4b89-8417-4bb333e05885 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lock "5b6e8205-003e-49c0-a73d-be2e032a8272" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.279s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.041626] env[62460]: DEBUG oslo_concurrency.lockutils [None req-87ee004f-61ca-441d-9f0d-e80687167704 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lock "5b6e8205-003e-49c0-a73d-be2e032a8272" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.056s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.041828] env[62460]: DEBUG nova.compute.manager [None req-87ee004f-61ca-441d-9f0d-e80687167704 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 931.045309] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a5a95d6-9ec3-4065-81df-9250ad8f5335 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.058632] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63829a3f-93b7-495f-8d86-56cc3833f0d5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.065681] env[62460]: DEBUG nova.compute.manager [None req-87ee004f-61ca-441d-9f0d-e80687167704 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62460) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 931.066355] env[62460]: DEBUG nova.objects.instance [None req-87ee004f-61ca-441d-9f0d-e80687167704 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lazy-loading 'flavor' on Instance uuid 5b6e8205-003e-49c0-a73d-be2e032a8272 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 931.076402] env[62460]: DEBUG nova.compute.manager [req-5d427f79-ca5c-46a9-9ceb-9387bdae23a5 req-0ed3cd3d-1c13-4886-8ab5-40c61b8a1341 service nova] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Detach interface failed, port_id=3304dbf2-549a-48d4-bd32-9fbd3d98cc3d, reason: Instance 61073a4e-c5e5-48f5-a43a-ce71bd18e48a could not be found. {{(pid=62460) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 931.151519] env[62460]: DEBUG oslo_vmware.api [None req-18074529-fe3b-4b5f-9546-b5cf5e11a600 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Task: {'id': task-1313875, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.187575} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.151519] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-18074529-fe3b-4b5f-9546-b5cf5e11a600 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 931.151519] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-18074529-fe3b-4b5f-9546-b5cf5e11a600 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Deleted contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 931.151519] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-18074529-fe3b-4b5f-9546-b5cf5e11a600 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 931.151519] env[62460]: INFO nova.compute.manager [None req-18074529-fe3b-4b5f-9546-b5cf5e11a600 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Took 1.11 seconds to destroy the instance on the hypervisor. [ 931.151519] env[62460]: DEBUG oslo.service.loopingcall [None req-18074529-fe3b-4b5f-9546-b5cf5e11a600 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 931.151860] env[62460]: DEBUG nova.compute.manager [-] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 931.151860] env[62460]: DEBUG nova.network.neutron [-] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 931.381894] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Releasing lock "refresh_cache-6a5934ed-30dc-4512-8cf2-1e4d9f20836c" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.382276] env[62460]: DEBUG nova.compute.manager [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] Instance network_info: |[{"id": "555d2fbe-16ac-4dd3-baa7-5b3b04b75670", "address": "fa:16:3e:f1:f5:35", "network": {"id": "de789887-9cf9-4614-a7e1-dab430aaabaf", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-656400207-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0da056d93bdf40c39d6e82e457727ff6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5116f690-f825-4fee-8a47-42b073e716c5", "external-id": "nsx-vlan-transportzone-692", "segmentation_id": 692, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap555d2fbe-16", "ovs_interfaceid": "555d2fbe-16ac-4dd3-baa7-5b3b04b75670", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 931.382809] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f1:f5:35', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5116f690-f825-4fee-8a47-42b073e716c5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '555d2fbe-16ac-4dd3-baa7-5b3b04b75670', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 931.390556] env[62460]: DEBUG oslo.service.loopingcall [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 931.390776] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 931.391010] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-29b7122d-2c3e-4f3f-97c3-e451bce982e2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.412739] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 931.412739] env[62460]: value = "task-1313877" [ 931.412739] env[62460]: _type = "Task" [ 931.412739] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.420679] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313877, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.423677] env[62460]: ERROR nova.scheduler.client.report [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [req-e3b040d0-0a20-4932-b221-83db4a931369] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 76e8f00d-65cc-4766-bb9d-5006fb51c728. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-e3b040d0-0a20-4932-b221-83db4a931369"}]} [ 931.442068] env[62460]: DEBUG nova.scheduler.client.report [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Refreshing inventories for resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 931.449416] env[62460]: DEBUG oslo_vmware.api [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313876, 'name': ReconfigVM_Task, 'duration_secs': 0.162642} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.449705] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Reconfigured VM instance instance-00000055 to detach disk 2000 {{(pid=62460) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 931.450734] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebfabf6e-ed9a-492c-a47c-678c452eef4e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.474148] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Reconfiguring VM instance instance-00000055 to attach disk [datastore2] 6b084ed4-e8f0-4063-bf2a-252740640753/6b084ed4-e8f0-4063-bf2a-252740640753.vmdk or device None with type thin {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 931.475253] env[62460]: DEBUG nova.scheduler.client.report [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Updating ProviderTree inventory for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 931.475456] env[62460]: DEBUG nova.compute.provider_tree [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 931.477470] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b3772174-60ea-43e0-bf1e-033825aaa14b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.495483] env[62460]: DEBUG oslo_vmware.api [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 931.495483] env[62460]: value = "task-1313878" [ 931.495483] env[62460]: _type = "Task" [ 931.495483] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.500161] env[62460]: DEBUG nova.scheduler.client.report [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Refreshing aggregate associations for resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728, aggregates: None {{(pid=62460) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 931.505574] env[62460]: DEBUG oslo_vmware.api [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313878, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.505954] env[62460]: INFO nova.compute.manager [-] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Took 1.28 seconds to deallocate network for instance. [ 931.521795] env[62460]: DEBUG nova.scheduler.client.report [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Refreshing trait associations for resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62460) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 931.572222] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-87ee004f-61ca-441d-9f0d-e80687167704 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 931.572528] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b57f5150-7bcf-4fce-8617-a498b8f78012 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.582652] env[62460]: DEBUG oslo_vmware.api [None req-87ee004f-61ca-441d-9f0d-e80687167704 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Waiting for the task: (returnval){ [ 931.582652] env[62460]: value = "task-1313879" [ 931.582652] env[62460]: _type = "Task" [ 931.582652] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.592276] env[62460]: DEBUG oslo_vmware.api [None req-87ee004f-61ca-441d-9f0d-e80687167704 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1313879, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.610265] env[62460]: DEBUG nova.compute.manager [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 931.613814] env[62460]: DEBUG nova.compute.manager [req-ed8799be-ef07-42dc-a361-91be644d2967 req-6460eca2-9988-4e4c-8174-d8cf67a8031b service nova] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] Received event network-changed-555d2fbe-16ac-4dd3-baa7-5b3b04b75670 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 931.613972] env[62460]: DEBUG nova.compute.manager [req-ed8799be-ef07-42dc-a361-91be644d2967 req-6460eca2-9988-4e4c-8174-d8cf67a8031b service nova] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] Refreshing instance network info cache due to event network-changed-555d2fbe-16ac-4dd3-baa7-5b3b04b75670. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 931.614144] env[62460]: DEBUG oslo_concurrency.lockutils [req-ed8799be-ef07-42dc-a361-91be644d2967 req-6460eca2-9988-4e4c-8174-d8cf67a8031b service nova] Acquiring lock "refresh_cache-6a5934ed-30dc-4512-8cf2-1e4d9f20836c" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.614562] env[62460]: DEBUG oslo_concurrency.lockutils [req-ed8799be-ef07-42dc-a361-91be644d2967 req-6460eca2-9988-4e4c-8174-d8cf67a8031b service nova] Acquired lock "refresh_cache-6a5934ed-30dc-4512-8cf2-1e4d9f20836c" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.614829] env[62460]: DEBUG nova.network.neutron [req-ed8799be-ef07-42dc-a361-91be644d2967 req-6460eca2-9988-4e4c-8174-d8cf67a8031b service nova] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] Refreshing network info cache for port 555d2fbe-16ac-4dd3-baa7-5b3b04b75670 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 931.648143] env[62460]: DEBUG nova.virt.hardware [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 931.648748] env[62460]: DEBUG nova.virt.hardware [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 931.648748] env[62460]: DEBUG nova.virt.hardware [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 931.648882] env[62460]: DEBUG nova.virt.hardware [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 931.648947] env[62460]: DEBUG nova.virt.hardware [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 931.649128] env[62460]: DEBUG nova.virt.hardware [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 931.649351] env[62460]: DEBUG nova.virt.hardware [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 931.649515] env[62460]: DEBUG nova.virt.hardware [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 931.649687] env[62460]: DEBUG nova.virt.hardware [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 931.649857] env[62460]: DEBUG nova.virt.hardware [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 931.650047] env[62460]: DEBUG nova.virt.hardware [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 931.650911] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c33799c8-25b3-49a0-9252-ab056bf7a1fb {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.661856] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e013e467-3132-463b-bc40-fd3f7ec5722b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.755913] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c00daac-3648-42da-a9b8-b7489b21e655 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.764242] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1282e9c-380c-4d14-a8c6-b67cc531bde7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.797439] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60f67195-c66e-4de5-82a3-c09fe1a4d874 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.805777] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b22c3e1-2e8a-4bda-ad84-44fd3fce9ce8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.821021] env[62460]: DEBUG nova.compute.provider_tree [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 931.922987] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313877, 'name': CreateVM_Task} progress is 99%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.004943] env[62460]: DEBUG oslo_vmware.api [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313878, 'name': ReconfigVM_Task, 'duration_secs': 0.26021} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.005246] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Reconfigured VM instance instance-00000055 to attach disk [datastore2] 6b084ed4-e8f0-4063-bf2a-252740640753/6b084ed4-e8f0-4063-bf2a-252740640753.vmdk or device None with type thin {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 932.005525] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Updating instance '6b084ed4-e8f0-4063-bf2a-252740640753' progress to 50 {{(pid=62460) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 932.012884] env[62460]: DEBUG oslo_concurrency.lockutils [None req-800eb51c-26fc-48d3-9ea6-3689387e9e69 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.092491] env[62460]: DEBUG oslo_vmware.api [None req-87ee004f-61ca-441d-9f0d-e80687167704 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1313879, 'name': PowerOffVM_Task, 'duration_secs': 0.23677} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.092765] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-87ee004f-61ca-441d-9f0d-e80687167704 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 932.093178] env[62460]: DEBUG nova.compute.manager [None req-87ee004f-61ca-441d-9f0d-e80687167704 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 932.093724] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a52d4cff-61e5-48e0-a2ff-1d955b3c8d4a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.125859] env[62460]: DEBUG nova.network.neutron [-] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.272443] env[62460]: DEBUG nova.network.neutron [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Successfully updated port: 2770c44d-3afd-4465-84ff-de06bb9b1a18 {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 932.361575] env[62460]: DEBUG nova.scheduler.client.report [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Updated inventory for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with generation 111 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 932.361901] env[62460]: DEBUG nova.compute.provider_tree [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Updating resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 generation from 111 to 112 during operation: update_inventory {{(pid=62460) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 932.362058] env[62460]: DEBUG nova.compute.provider_tree [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 932.365967] env[62460]: DEBUG nova.network.neutron [req-ed8799be-ef07-42dc-a361-91be644d2967 req-6460eca2-9988-4e4c-8174-d8cf67a8031b service nova] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] Updated VIF entry in instance network info cache for port 555d2fbe-16ac-4dd3-baa7-5b3b04b75670. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 932.366335] env[62460]: DEBUG nova.network.neutron [req-ed8799be-ef07-42dc-a361-91be644d2967 req-6460eca2-9988-4e4c-8174-d8cf67a8031b service nova] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] Updating instance_info_cache with network_info: [{"id": "555d2fbe-16ac-4dd3-baa7-5b3b04b75670", "address": "fa:16:3e:f1:f5:35", "network": {"id": "de789887-9cf9-4614-a7e1-dab430aaabaf", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-656400207-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0da056d93bdf40c39d6e82e457727ff6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5116f690-f825-4fee-8a47-42b073e716c5", "external-id": "nsx-vlan-transportzone-692", "segmentation_id": 692, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap555d2fbe-16", "ovs_interfaceid": "555d2fbe-16ac-4dd3-baa7-5b3b04b75670", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.424323] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313877, 'name': CreateVM_Task} progress is 99%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.512512] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98c02208-0dad-49e5-9fb0-af58406973b7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.531097] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f65a1145-3ffd-49e8-b21a-2652cc30942b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.549735] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Updating instance '6b084ed4-e8f0-4063-bf2a-252740640753' progress to 67 {{(pid=62460) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 932.554558] env[62460]: DEBUG nova.compute.manager [req-141e61ea-073a-42f1-88c0-715530022ffe req-a3c43281-cece-4a53-9486-82cb3d56dea2 service nova] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Received event network-vif-deleted-530faf0d-567e-464f-99de-5f2f1d7afaf2 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 932.605355] env[62460]: DEBUG oslo_concurrency.lockutils [None req-87ee004f-61ca-441d-9f0d-e80687167704 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lock "5b6e8205-003e-49c0-a73d-be2e032a8272" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.564s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.629829] env[62460]: INFO nova.compute.manager [-] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Took 1.48 seconds to deallocate network for instance. [ 932.776431] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Acquiring lock "refresh_cache-9c63946c-f294-42ea-979e-68db26d88858" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.776589] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Acquired lock "refresh_cache-9c63946c-f294-42ea-979e-68db26d88858" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.776751] env[62460]: DEBUG nova.network.neutron [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 932.869114] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.273s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.869792] env[62460]: DEBUG nova.compute.manager [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 932.872895] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.518s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.875026] env[62460]: INFO nova.compute.claims [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 932.877366] env[62460]: DEBUG oslo_concurrency.lockutils [req-ed8799be-ef07-42dc-a361-91be644d2967 req-6460eca2-9988-4e4c-8174-d8cf67a8031b service nova] Releasing lock "refresh_cache-6a5934ed-30dc-4512-8cf2-1e4d9f20836c" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.924907] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313877, 'name': CreateVM_Task, 'duration_secs': 1.307699} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.924907] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 932.925477] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.925733] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.926254] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 932.926552] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-be2abce0-0d10-43b8-bdbd-e799e97f79f3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.931263] env[62460]: DEBUG oslo_vmware.api [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 932.931263] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]524db1bd-e7cd-00cc-86fc-9779236c66a3" [ 932.931263] env[62460]: _type = "Task" [ 932.931263] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.939239] env[62460]: DEBUG oslo_vmware.api [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]524db1bd-e7cd-00cc-86fc-9779236c66a3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.090896] env[62460]: DEBUG nova.network.neutron [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Port 3eaf04a4-feda-4a00-908e-d121fc775ecb binding to destination host cpu-1 is already ACTIVE {{(pid=62460) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 933.135678] env[62460]: DEBUG oslo_concurrency.lockutils [None req-18074529-fe3b-4b5f-9546-b5cf5e11a600 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.307177] env[62460]: DEBUG nova.network.neutron [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 933.374840] env[62460]: DEBUG nova.compute.utils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 933.376216] env[62460]: DEBUG nova.compute.manager [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 933.376368] env[62460]: DEBUG nova.network.neutron [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 933.416241] env[62460]: DEBUG nova.policy [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '13ee01926a10432dbb135f08d7924da3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '28e73bd27f474a81a0aa7d4bb2d8efb4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 933.441791] env[62460]: DEBUG oslo_vmware.api [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]524db1bd-e7cd-00cc-86fc-9779236c66a3, 'name': SearchDatastore_Task, 'duration_secs': 0.039556} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.441947] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 933.442206] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 933.442449] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.442603] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.442795] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 933.443086] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5a0c2b5e-c729-49ee-a316-c8265607c906 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.450965] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 933.451298] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 933.451873] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-08e62837-b43d-4579-ad36-40b3b2e5233d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.459267] env[62460]: DEBUG oslo_vmware.api [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 933.459267] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]528adabd-017d-585e-7cd4-6d583c24f18f" [ 933.459267] env[62460]: _type = "Task" [ 933.459267] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.460114] env[62460]: DEBUG nova.network.neutron [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Updating instance_info_cache with network_info: [{"id": "2770c44d-3afd-4465-84ff-de06bb9b1a18", "address": "fa:16:3e:83:3c:6b", "network": {"id": "0c9ffa5f-8a93-402d-bb98-8d88ef381c15", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-919853047-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28e73bd27f474a81a0aa7d4bb2d8efb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "450939f7-f74b-41f7-93f7-b4fde6a6fbed", "external-id": "nsx-vlan-transportzone-866", "segmentation_id": 866, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2770c44d-3a", "ovs_interfaceid": "2770c44d-3afd-4465-84ff-de06bb9b1a18", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.472422] env[62460]: DEBUG oslo_vmware.api [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]528adabd-017d-585e-7cd4-6d583c24f18f, 'name': SearchDatastore_Task, 'duration_secs': 0.008332} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.473192] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c5207acc-36f9-411e-bf18-d7608c682bf2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.478629] env[62460]: DEBUG oslo_vmware.api [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 933.478629] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]5268117c-31ed-8410-9840-b26dae43ce3d" [ 933.478629] env[62460]: _type = "Task" [ 933.478629] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.486095] env[62460]: DEBUG oslo_vmware.api [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5268117c-31ed-8410-9840-b26dae43ce3d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.537948] env[62460]: DEBUG nova.objects.instance [None req-d0227866-9552-40d6-ab39-e5666c57a6b5 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lazy-loading 'flavor' on Instance uuid 5b6e8205-003e-49c0-a73d-be2e032a8272 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 933.639057] env[62460]: DEBUG nova.compute.manager [req-448cc0d7-302b-4e47-a996-312375f4ac11 req-63cdd58d-fa66-4917-8e90-eec037d68a4f service nova] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Received event network-vif-plugged-2770c44d-3afd-4465-84ff-de06bb9b1a18 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 933.639057] env[62460]: DEBUG oslo_concurrency.lockutils [req-448cc0d7-302b-4e47-a996-312375f4ac11 req-63cdd58d-fa66-4917-8e90-eec037d68a4f service nova] Acquiring lock "9c63946c-f294-42ea-979e-68db26d88858-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.639057] env[62460]: DEBUG oslo_concurrency.lockutils [req-448cc0d7-302b-4e47-a996-312375f4ac11 req-63cdd58d-fa66-4917-8e90-eec037d68a4f service nova] Lock "9c63946c-f294-42ea-979e-68db26d88858-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.639366] env[62460]: DEBUG oslo_concurrency.lockutils [req-448cc0d7-302b-4e47-a996-312375f4ac11 req-63cdd58d-fa66-4917-8e90-eec037d68a4f service nova] Lock "9c63946c-f294-42ea-979e-68db26d88858-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.639366] env[62460]: DEBUG nova.compute.manager [req-448cc0d7-302b-4e47-a996-312375f4ac11 req-63cdd58d-fa66-4917-8e90-eec037d68a4f service nova] [instance: 9c63946c-f294-42ea-979e-68db26d88858] No waiting events found dispatching network-vif-plugged-2770c44d-3afd-4465-84ff-de06bb9b1a18 {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 933.639575] env[62460]: WARNING nova.compute.manager [req-448cc0d7-302b-4e47-a996-312375f4ac11 req-63cdd58d-fa66-4917-8e90-eec037d68a4f service nova] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Received unexpected event network-vif-plugged-2770c44d-3afd-4465-84ff-de06bb9b1a18 for instance with vm_state building and task_state spawning. [ 933.639659] env[62460]: DEBUG nova.compute.manager [req-448cc0d7-302b-4e47-a996-312375f4ac11 req-63cdd58d-fa66-4917-8e90-eec037d68a4f service nova] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Received event network-changed-2770c44d-3afd-4465-84ff-de06bb9b1a18 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 933.642958] env[62460]: DEBUG nova.compute.manager [req-448cc0d7-302b-4e47-a996-312375f4ac11 req-63cdd58d-fa66-4917-8e90-eec037d68a4f service nova] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Refreshing instance network info cache due to event network-changed-2770c44d-3afd-4465-84ff-de06bb9b1a18. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 933.642958] env[62460]: DEBUG oslo_concurrency.lockutils [req-448cc0d7-302b-4e47-a996-312375f4ac11 req-63cdd58d-fa66-4917-8e90-eec037d68a4f service nova] Acquiring lock "refresh_cache-9c63946c-f294-42ea-979e-68db26d88858" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.712140] env[62460]: DEBUG nova.network.neutron [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Successfully created port: 73df6556-a3c8-4ee6-9881-8b978f97ca6f {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 933.879819] env[62460]: DEBUG nova.compute.manager [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 933.965690] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Releasing lock "refresh_cache-9c63946c-f294-42ea-979e-68db26d88858" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 933.966792] env[62460]: DEBUG nova.compute.manager [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Instance network_info: |[{"id": "2770c44d-3afd-4465-84ff-de06bb9b1a18", "address": "fa:16:3e:83:3c:6b", "network": {"id": "0c9ffa5f-8a93-402d-bb98-8d88ef381c15", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-919853047-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28e73bd27f474a81a0aa7d4bb2d8efb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "450939f7-f74b-41f7-93f7-b4fde6a6fbed", "external-id": "nsx-vlan-transportzone-866", "segmentation_id": 866, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2770c44d-3a", "ovs_interfaceid": "2770c44d-3afd-4465-84ff-de06bb9b1a18", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 933.966792] env[62460]: DEBUG oslo_concurrency.lockutils [req-448cc0d7-302b-4e47-a996-312375f4ac11 req-63cdd58d-fa66-4917-8e90-eec037d68a4f service nova] Acquired lock "refresh_cache-9c63946c-f294-42ea-979e-68db26d88858" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.966792] env[62460]: DEBUG nova.network.neutron [req-448cc0d7-302b-4e47-a996-312375f4ac11 req-63cdd58d-fa66-4917-8e90-eec037d68a4f service nova] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Refreshing network info cache for port 2770c44d-3afd-4465-84ff-de06bb9b1a18 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 933.968550] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:83:3c:6b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '450939f7-f74b-41f7-93f7-b4fde6a6fbed', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2770c44d-3afd-4465-84ff-de06bb9b1a18', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 933.975508] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Creating folder: Project (28e73bd27f474a81a0aa7d4bb2d8efb4). Parent ref: group-v281134. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 933.980669] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6c43d3ce-9c15-46d1-812a-6433cdd3d507 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.992889] env[62460]: DEBUG oslo_vmware.api [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5268117c-31ed-8410-9840-b26dae43ce3d, 'name': SearchDatastore_Task, 'duration_secs': 0.008775} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.993222] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 933.993505] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 6a5934ed-30dc-4512-8cf2-1e4d9f20836c/6a5934ed-30dc-4512-8cf2-1e4d9f20836c.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 933.996314] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-946667e2-0e5f-41bd-9408-9636e874d404 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.000260] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Created folder: Project (28e73bd27f474a81a0aa7d4bb2d8efb4) in parent group-v281134. [ 934.000260] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Creating folder: Instances. Parent ref: group-v281246. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 934.000729] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ddea9f08-ca6a-4612-9d37-d627f0515106 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.006026] env[62460]: DEBUG oslo_vmware.api [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 934.006026] env[62460]: value = "task-1313881" [ 934.006026] env[62460]: _type = "Task" [ 934.006026] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.010211] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Created folder: Instances in parent group-v281246. [ 934.010573] env[62460]: DEBUG oslo.service.loopingcall [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 934.016135] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 934.016436] env[62460]: DEBUG oslo_vmware.api [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313881, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.016852] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-03b394d6-aef0-4098-9889-e669d4ad91a1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.044473] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d0227866-9552-40d6-ab39-e5666c57a6b5 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Acquiring lock "refresh_cache-5b6e8205-003e-49c0-a73d-be2e032a8272" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.044564] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d0227866-9552-40d6-ab39-e5666c57a6b5 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Acquired lock "refresh_cache-5b6e8205-003e-49c0-a73d-be2e032a8272" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.045201] env[62460]: DEBUG nova.network.neutron [None req-d0227866-9552-40d6-ab39-e5666c57a6b5 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 934.045201] env[62460]: DEBUG nova.objects.instance [None req-d0227866-9552-40d6-ab39-e5666c57a6b5 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lazy-loading 'info_cache' on Instance uuid 5b6e8205-003e-49c0-a73d-be2e032a8272 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 934.048152] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 934.048152] env[62460]: value = "task-1313883" [ 934.048152] env[62460]: _type = "Task" [ 934.048152] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.058073] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313883, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.113914] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c840fec6-066e-44b9-9b53-4cce15304b21 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.125275] env[62460]: DEBUG oslo_concurrency.lockutils [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "6b084ed4-e8f0-4063-bf2a-252740640753-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.125575] env[62460]: DEBUG oslo_concurrency.lockutils [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "6b084ed4-e8f0-4063-bf2a-252740640753-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.125778] env[62460]: DEBUG oslo_concurrency.lockutils [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "6b084ed4-e8f0-4063-bf2a-252740640753-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.134486] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1570752-32bf-4888-be79-1bef98bcc5b7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.168189] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3ef323e-39c7-4faf-83f5-d46f16877d8d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.177354] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4d3692f-925a-43ab-a72f-2ca308c26041 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.192398] env[62460]: DEBUG nova.compute.provider_tree [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 934.289086] env[62460]: DEBUG nova.network.neutron [req-448cc0d7-302b-4e47-a996-312375f4ac11 req-63cdd58d-fa66-4917-8e90-eec037d68a4f service nova] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Updated VIF entry in instance network info cache for port 2770c44d-3afd-4465-84ff-de06bb9b1a18. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 934.289728] env[62460]: DEBUG nova.network.neutron [req-448cc0d7-302b-4e47-a996-312375f4ac11 req-63cdd58d-fa66-4917-8e90-eec037d68a4f service nova] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Updating instance_info_cache with network_info: [{"id": "2770c44d-3afd-4465-84ff-de06bb9b1a18", "address": "fa:16:3e:83:3c:6b", "network": {"id": "0c9ffa5f-8a93-402d-bb98-8d88ef381c15", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-919853047-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28e73bd27f474a81a0aa7d4bb2d8efb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "450939f7-f74b-41f7-93f7-b4fde6a6fbed", "external-id": "nsx-vlan-transportzone-866", "segmentation_id": 866, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2770c44d-3a", "ovs_interfaceid": "2770c44d-3afd-4465-84ff-de06bb9b1a18", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 934.514024] env[62460]: DEBUG oslo_vmware.api [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313881, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.548800] env[62460]: DEBUG nova.objects.base [None req-d0227866-9552-40d6-ab39-e5666c57a6b5 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Object Instance<5b6e8205-003e-49c0-a73d-be2e032a8272> lazy-loaded attributes: flavor,info_cache {{(pid=62460) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 934.559132] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313883, 'name': CreateVM_Task} progress is 25%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.695603] env[62460]: DEBUG nova.scheduler.client.report [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 934.792431] env[62460]: DEBUG oslo_concurrency.lockutils [req-448cc0d7-302b-4e47-a996-312375f4ac11 req-63cdd58d-fa66-4917-8e90-eec037d68a4f service nova] Releasing lock "refresh_cache-9c63946c-f294-42ea-979e-68db26d88858" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.891509] env[62460]: DEBUG nova.compute.manager [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 934.916077] env[62460]: DEBUG nova.virt.hardware [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 934.916414] env[62460]: DEBUG nova.virt.hardware [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 934.916586] env[62460]: DEBUG nova.virt.hardware [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 934.916774] env[62460]: DEBUG nova.virt.hardware [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 934.916927] env[62460]: DEBUG nova.virt.hardware [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 934.917098] env[62460]: DEBUG nova.virt.hardware [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 934.917321] env[62460]: DEBUG nova.virt.hardware [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 934.917494] env[62460]: DEBUG nova.virt.hardware [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 934.917669] env[62460]: DEBUG nova.virt.hardware [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 934.917837] env[62460]: DEBUG nova.virt.hardware [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 934.918061] env[62460]: DEBUG nova.virt.hardware [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 934.919191] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ef5c180-4e1e-4620-bd59-7655aedf07d6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.927174] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b088677e-93e1-481d-8165-61bc4c3e563d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.016931] env[62460]: DEBUG oslo_vmware.api [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313881, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.549663} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.017322] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 6a5934ed-30dc-4512-8cf2-1e4d9f20836c/6a5934ed-30dc-4512-8cf2-1e4d9f20836c.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 935.017635] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 935.017962] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-12c93afa-22f1-4722-806a-e5812d4ba286 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.026087] env[62460]: DEBUG oslo_vmware.api [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 935.026087] env[62460]: value = "task-1313884" [ 935.026087] env[62460]: _type = "Task" [ 935.026087] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.036796] env[62460]: DEBUG oslo_vmware.api [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313884, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.064712] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313883, 'name': CreateVM_Task} progress is 99%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.202882] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.330s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.203622] env[62460]: DEBUG nova.compute.manager [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 935.209728] env[62460]: DEBUG oslo_concurrency.lockutils [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.295s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.211284] env[62460]: INFO nova.compute.claims [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 935.219243] env[62460]: DEBUG nova.network.neutron [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Successfully updated port: 73df6556-a3c8-4ee6-9881-8b978f97ca6f {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 935.225465] env[62460]: DEBUG oslo_concurrency.lockutils [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "refresh_cache-6b084ed4-e8f0-4063-bf2a-252740640753" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.225650] env[62460]: DEBUG oslo_concurrency.lockutils [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquired lock "refresh_cache-6b084ed4-e8f0-4063-bf2a-252740640753" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.225850] env[62460]: DEBUG nova.network.neutron [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 935.373331] env[62460]: DEBUG nova.network.neutron [None req-d0227866-9552-40d6-ab39-e5666c57a6b5 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Updating instance_info_cache with network_info: [{"id": "3515dec2-bd6c-4f96-aff1-8dd32de6f7d6", "address": "fa:16:3e:53:9c:d5", "network": {"id": "4cb03486-3bea-4c6a-a019-37486f271107", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-2094026403-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.174", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfc48e301fb04231b88ec28a062bda4e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c7d2575f-b92f-44ec-a863-634cb76631a2", "external-id": "nsx-vlan-transportzone-794", "segmentation_id": 794, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3515dec2-bd", "ovs_interfaceid": "3515dec2-bd6c-4f96-aff1-8dd32de6f7d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.536276] env[62460]: DEBUG oslo_vmware.api [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313884, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069283} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.536628] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 935.537402] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79c20c6b-6adf-4518-b34a-7173d355e7ef {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.559059] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] Reconfiguring VM instance instance-00000059 to attach disk [datastore1] 6a5934ed-30dc-4512-8cf2-1e4d9f20836c/6a5934ed-30dc-4512-8cf2-1e4d9f20836c.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 935.562052] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1f407641-56d9-4fb1-813d-e22fec4f4623 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.582379] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313883, 'name': CreateVM_Task} progress is 99%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.583582] env[62460]: DEBUG oslo_vmware.api [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 935.583582] env[62460]: value = "task-1313885" [ 935.583582] env[62460]: _type = "Task" [ 935.583582] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.590771] env[62460]: DEBUG oslo_vmware.api [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313885, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.666794] env[62460]: DEBUG nova.compute.manager [req-5e2e0176-423b-410f-8a88-40b01b3d6eaa req-05e0a3c8-36ad-4e8b-80f7-13f0e6265574 service nova] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Received event network-vif-plugged-73df6556-a3c8-4ee6-9881-8b978f97ca6f {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 935.667158] env[62460]: DEBUG oslo_concurrency.lockutils [req-5e2e0176-423b-410f-8a88-40b01b3d6eaa req-05e0a3c8-36ad-4e8b-80f7-13f0e6265574 service nova] Acquiring lock "003191d3-fe38-4bde-8d7c-75b4aad413b4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.667265] env[62460]: DEBUG oslo_concurrency.lockutils [req-5e2e0176-423b-410f-8a88-40b01b3d6eaa req-05e0a3c8-36ad-4e8b-80f7-13f0e6265574 service nova] Lock "003191d3-fe38-4bde-8d7c-75b4aad413b4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.667458] env[62460]: DEBUG oslo_concurrency.lockutils [req-5e2e0176-423b-410f-8a88-40b01b3d6eaa req-05e0a3c8-36ad-4e8b-80f7-13f0e6265574 service nova] Lock "003191d3-fe38-4bde-8d7c-75b4aad413b4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.667649] env[62460]: DEBUG nova.compute.manager [req-5e2e0176-423b-410f-8a88-40b01b3d6eaa req-05e0a3c8-36ad-4e8b-80f7-13f0e6265574 service nova] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] No waiting events found dispatching network-vif-plugged-73df6556-a3c8-4ee6-9881-8b978f97ca6f {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 935.667834] env[62460]: WARNING nova.compute.manager [req-5e2e0176-423b-410f-8a88-40b01b3d6eaa req-05e0a3c8-36ad-4e8b-80f7-13f0e6265574 service nova] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Received unexpected event network-vif-plugged-73df6556-a3c8-4ee6-9881-8b978f97ca6f for instance with vm_state building and task_state spawning. [ 935.668024] env[62460]: DEBUG nova.compute.manager [req-5e2e0176-423b-410f-8a88-40b01b3d6eaa req-05e0a3c8-36ad-4e8b-80f7-13f0e6265574 service nova] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Received event network-changed-73df6556-a3c8-4ee6-9881-8b978f97ca6f {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 935.668248] env[62460]: DEBUG nova.compute.manager [req-5e2e0176-423b-410f-8a88-40b01b3d6eaa req-05e0a3c8-36ad-4e8b-80f7-13f0e6265574 service nova] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Refreshing instance network info cache due to event network-changed-73df6556-a3c8-4ee6-9881-8b978f97ca6f. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 935.668463] env[62460]: DEBUG oslo_concurrency.lockutils [req-5e2e0176-423b-410f-8a88-40b01b3d6eaa req-05e0a3c8-36ad-4e8b-80f7-13f0e6265574 service nova] Acquiring lock "refresh_cache-003191d3-fe38-4bde-8d7c-75b4aad413b4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.668613] env[62460]: DEBUG oslo_concurrency.lockutils [req-5e2e0176-423b-410f-8a88-40b01b3d6eaa req-05e0a3c8-36ad-4e8b-80f7-13f0e6265574 service nova] Acquired lock "refresh_cache-003191d3-fe38-4bde-8d7c-75b4aad413b4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.668801] env[62460]: DEBUG nova.network.neutron [req-5e2e0176-423b-410f-8a88-40b01b3d6eaa req-05e0a3c8-36ad-4e8b-80f7-13f0e6265574 service nova] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Refreshing network info cache for port 73df6556-a3c8-4ee6-9881-8b978f97ca6f {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 935.716226] env[62460]: DEBUG nova.compute.utils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 935.720282] env[62460]: DEBUG nova.compute.manager [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 935.720466] env[62460]: DEBUG nova.network.neutron [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 935.723601] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Acquiring lock "refresh_cache-003191d3-fe38-4bde-8d7c-75b4aad413b4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.813151] env[62460]: DEBUG nova.policy [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '13ee01926a10432dbb135f08d7924da3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '28e73bd27f474a81a0aa7d4bb2d8efb4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 935.876231] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d0227866-9552-40d6-ab39-e5666c57a6b5 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Releasing lock "refresh_cache-5b6e8205-003e-49c0-a73d-be2e032a8272" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.065593] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313883, 'name': CreateVM_Task, 'duration_secs': 1.705843} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.065819] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 936.066542] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.066719] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.067054] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 936.067313] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-468b7dea-c99c-40c3-a088-349d08a19147 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.071532] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Waiting for the task: (returnval){ [ 936.071532] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]526850fe-e72b-1fca-bb65-3dda0316f4e9" [ 936.071532] env[62460]: _type = "Task" [ 936.071532] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.081908] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]526850fe-e72b-1fca-bb65-3dda0316f4e9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.093267] env[62460]: DEBUG oslo_vmware.api [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313885, 'name': ReconfigVM_Task, 'duration_secs': 0.274426} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.093267] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] Reconfigured VM instance instance-00000059 to attach disk [datastore1] 6a5934ed-30dc-4512-8cf2-1e4d9f20836c/6a5934ed-30dc-4512-8cf2-1e4d9f20836c.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 936.093267] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-07c5be1d-c70a-4614-ae41-41d67d9dae49 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.102028] env[62460]: DEBUG oslo_vmware.api [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 936.102028] env[62460]: value = "task-1313886" [ 936.102028] env[62460]: _type = "Task" [ 936.102028] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.107222] env[62460]: DEBUG oslo_vmware.api [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313886, 'name': Rename_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.200872] env[62460]: DEBUG nova.network.neutron [req-5e2e0176-423b-410f-8a88-40b01b3d6eaa req-05e0a3c8-36ad-4e8b-80f7-13f0e6265574 service nova] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 936.223985] env[62460]: DEBUG nova.compute.manager [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 936.311092] env[62460]: DEBUG nova.network.neutron [req-5e2e0176-423b-410f-8a88-40b01b3d6eaa req-05e0a3c8-36ad-4e8b-80f7-13f0e6265574 service nova] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.315056] env[62460]: DEBUG nova.network.neutron [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Updating instance_info_cache with network_info: [{"id": "3eaf04a4-feda-4a00-908e-d121fc775ecb", "address": "fa:16:3e:98:27:28", "network": {"id": "3c49720c-94c9-4e2f-bbb7-4266cc97faee", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1650281790-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfa5a65b0f614e769de5b3aa77bf869e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3eaf04a4-fe", "ovs_interfaceid": "3eaf04a4-feda-4a00-908e-d121fc775ecb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.380202] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0227866-9552-40d6-ab39-e5666c57a6b5 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 936.381388] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cf10b3b4-3841-4ab4-a55f-b27fb10c34a9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.388056] env[62460]: DEBUG oslo_vmware.api [None req-d0227866-9552-40d6-ab39-e5666c57a6b5 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Waiting for the task: (returnval){ [ 936.388056] env[62460]: value = "task-1313887" [ 936.388056] env[62460]: _type = "Task" [ 936.388056] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.398698] env[62460]: DEBUG oslo_vmware.api [None req-d0227866-9552-40d6-ab39-e5666c57a6b5 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1313887, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.438918] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76619ed9-2e62-4873-bed6-88ef7c194b2d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.448216] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b96bdb61-092d-44e8-b7ed-9dc876fbd677 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.480599] env[62460]: DEBUG nova.network.neutron [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Successfully created port: 96b72a46-8d24-420d-a0f2-407651530be5 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 936.483174] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-125b642e-9cde-4ae0-8021-3c58b6605e81 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.490897] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f132f4c1-af2e-4e0d-a983-b3bcbc9d77e9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.505675] env[62460]: DEBUG nova.compute.provider_tree [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 936.582996] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]526850fe-e72b-1fca-bb65-3dda0316f4e9, 'name': SearchDatastore_Task, 'duration_secs': 0.009087} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.583506] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.583907] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 936.584287] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.584562] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.584853] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 936.585255] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-70c198d3-1151-4022-88b6-0d0279447246 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.600061] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 936.600061] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 936.600061] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ff008a60-8e4b-4572-adda-694ac86f3e5d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.606887] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Waiting for the task: (returnval){ [ 936.606887] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52e5cb08-d530-a424-c583-2449819bdc52" [ 936.606887] env[62460]: _type = "Task" [ 936.606887] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.610389] env[62460]: DEBUG oslo_vmware.api [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313886, 'name': Rename_Task, 'duration_secs': 0.142245} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.613509] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 936.613941] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a5ed7bf7-99aa-4ae1-8daa-ee523dccef9c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.620802] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52e5cb08-d530-a424-c583-2449819bdc52, 'name': SearchDatastore_Task, 'duration_secs': 0.00914} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.622664] env[62460]: DEBUG oslo_vmware.api [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 936.622664] env[62460]: value = "task-1313888" [ 936.622664] env[62460]: _type = "Task" [ 936.622664] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.623140] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3fc65a09-ebab-4a5d-8b4a-4b02ac5afb5b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.633146] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Waiting for the task: (returnval){ [ 936.633146] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52f59a81-9a81-a76f-6c2f-829394a02b97" [ 936.633146] env[62460]: _type = "Task" [ 936.633146] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.639242] env[62460]: DEBUG oslo_vmware.api [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313888, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.650451] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52f59a81-9a81-a76f-6c2f-829394a02b97, 'name': SearchDatastore_Task, 'duration_secs': 0.014248} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.653762] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.653762] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 9c63946c-f294-42ea-979e-68db26d88858/9c63946c-f294-42ea-979e-68db26d88858.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 936.653762] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7d2f83e5-162b-4c1f-b7f8-bc97f42ef329 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.660021] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Waiting for the task: (returnval){ [ 936.660021] env[62460]: value = "task-1313889" [ 936.660021] env[62460]: _type = "Task" [ 936.660021] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.665084] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313889, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.818293] env[62460]: DEBUG oslo_concurrency.lockutils [req-5e2e0176-423b-410f-8a88-40b01b3d6eaa req-05e0a3c8-36ad-4e8b-80f7-13f0e6265574 service nova] Releasing lock "refresh_cache-003191d3-fe38-4bde-8d7c-75b4aad413b4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.818665] env[62460]: DEBUG oslo_concurrency.lockutils [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Releasing lock "refresh_cache-6b084ed4-e8f0-4063-bf2a-252740640753" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.822431] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Acquired lock "refresh_cache-003191d3-fe38-4bde-8d7c-75b4aad413b4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.822594] env[62460]: DEBUG nova.network.neutron [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 936.898178] env[62460]: DEBUG oslo_vmware.api [None req-d0227866-9552-40d6-ab39-e5666c57a6b5 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1313887, 'name': PowerOnVM_Task, 'duration_secs': 0.382204} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.898579] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0227866-9552-40d6-ab39-e5666c57a6b5 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 936.898780] env[62460]: DEBUG nova.compute.manager [None req-d0227866-9552-40d6-ab39-e5666c57a6b5 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 936.899681] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d11caef0-0a2e-4f35-be2b-bf554657ac00 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.009038] env[62460]: DEBUG nova.scheduler.client.report [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 937.136567] env[62460]: DEBUG oslo_vmware.api [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313888, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.167957] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313889, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.237612] env[62460]: DEBUG nova.compute.manager [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 937.266019] env[62460]: DEBUG nova.virt.hardware [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 937.266019] env[62460]: DEBUG nova.virt.hardware [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 937.266019] env[62460]: DEBUG nova.virt.hardware [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 937.266019] env[62460]: DEBUG nova.virt.hardware [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 937.266019] env[62460]: DEBUG nova.virt.hardware [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 937.266019] env[62460]: DEBUG nova.virt.hardware [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 937.266598] env[62460]: DEBUG nova.virt.hardware [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 937.266962] env[62460]: DEBUG nova.virt.hardware [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 937.267387] env[62460]: DEBUG nova.virt.hardware [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 937.267803] env[62460]: DEBUG nova.virt.hardware [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 937.268325] env[62460]: DEBUG nova.virt.hardware [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 937.269585] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0f8e5dc-9012-47f8-8284-d441ecb08fff {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.278832] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d82388bb-184b-406a-b518-f2e84d685227 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.345551] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0f8dde3-3d55-4991-96f8-767fb451c20d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.364878] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dea00c5-e846-41b1-8690-05d5da89b9cd {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.371846] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Updating instance '6b084ed4-e8f0-4063-bf2a-252740640753' progress to 83 {{(pid=62460) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 937.375981] env[62460]: DEBUG nova.network.neutron [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 937.514090] env[62460]: DEBUG oslo_concurrency.lockutils [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.308s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.514687] env[62460]: DEBUG nova.compute.manager [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 937.518787] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.970s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.519097] env[62460]: DEBUG nova.objects.instance [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lazy-loading 'pci_requests' on Instance uuid 896f8ff4-e45f-4403-a727-03ee25e58609 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 937.531999] env[62460]: DEBUG nova.network.neutron [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Updating instance_info_cache with network_info: [{"id": "73df6556-a3c8-4ee6-9881-8b978f97ca6f", "address": "fa:16:3e:2b:5b:2e", "network": {"id": "0c9ffa5f-8a93-402d-bb98-8d88ef381c15", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-919853047-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28e73bd27f474a81a0aa7d4bb2d8efb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "450939f7-f74b-41f7-93f7-b4fde6a6fbed", "external-id": "nsx-vlan-transportzone-866", "segmentation_id": 866, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap73df6556-a3", "ovs_interfaceid": "73df6556-a3c8-4ee6-9881-8b978f97ca6f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 937.637037] env[62460]: DEBUG oslo_vmware.api [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313888, 'name': PowerOnVM_Task, 'duration_secs': 0.68538} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.637320] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 937.637532] env[62460]: INFO nova.compute.manager [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] Took 8.39 seconds to spawn the instance on the hypervisor. [ 937.637718] env[62460]: DEBUG nova.compute.manager [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 937.638769] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8983e734-6b97-4d0a-80f6-e98cb903bdba {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.667589] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313889, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.538199} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.667589] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 9c63946c-f294-42ea-979e-68db26d88858/9c63946c-f294-42ea-979e-68db26d88858.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 937.667838] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 937.668565] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e3d56970-6475-473c-9f06-3a181e65e2f1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.674929] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Waiting for the task: (returnval){ [ 937.674929] env[62460]: value = "task-1313890" [ 937.674929] env[62460]: _type = "Task" [ 937.674929] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.683637] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313890, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.881198] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 937.881628] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f95d1163-ace8-49df-998b-8e411eb25a9b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.889962] env[62460]: DEBUG oslo_vmware.api [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 937.889962] env[62460]: value = "task-1313891" [ 937.889962] env[62460]: _type = "Task" [ 937.889962] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.900458] env[62460]: DEBUG oslo_vmware.api [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313891, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.962808] env[62460]: DEBUG nova.compute.manager [req-5100d0a3-58e8-47e4-9bac-b95da6b200ab req-b52353a4-fbd8-445e-ad3a-75983ab9c594 service nova] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Received event network-vif-plugged-96b72a46-8d24-420d-a0f2-407651530be5 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 937.962899] env[62460]: DEBUG oslo_concurrency.lockutils [req-5100d0a3-58e8-47e4-9bac-b95da6b200ab req-b52353a4-fbd8-445e-ad3a-75983ab9c594 service nova] Acquiring lock "db331001-1744-4c42-a40c-87f3e60da4b8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.963151] env[62460]: DEBUG oslo_concurrency.lockutils [req-5100d0a3-58e8-47e4-9bac-b95da6b200ab req-b52353a4-fbd8-445e-ad3a-75983ab9c594 service nova] Lock "db331001-1744-4c42-a40c-87f3e60da4b8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.963294] env[62460]: DEBUG oslo_concurrency.lockutils [req-5100d0a3-58e8-47e4-9bac-b95da6b200ab req-b52353a4-fbd8-445e-ad3a-75983ab9c594 service nova] Lock "db331001-1744-4c42-a40c-87f3e60da4b8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.963475] env[62460]: DEBUG nova.compute.manager [req-5100d0a3-58e8-47e4-9bac-b95da6b200ab req-b52353a4-fbd8-445e-ad3a-75983ab9c594 service nova] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] No waiting events found dispatching network-vif-plugged-96b72a46-8d24-420d-a0f2-407651530be5 {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 937.963645] env[62460]: WARNING nova.compute.manager [req-5100d0a3-58e8-47e4-9bac-b95da6b200ab req-b52353a4-fbd8-445e-ad3a-75983ab9c594 service nova] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Received unexpected event network-vif-plugged-96b72a46-8d24-420d-a0f2-407651530be5 for instance with vm_state building and task_state spawning. [ 938.020452] env[62460]: DEBUG nova.compute.utils [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 938.021860] env[62460]: DEBUG nova.compute.manager [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 938.026090] env[62460]: DEBUG nova.network.neutron [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 938.030051] env[62460]: DEBUG nova.objects.instance [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lazy-loading 'numa_topology' on Instance uuid 896f8ff4-e45f-4403-a727-03ee25e58609 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 938.034579] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Releasing lock "refresh_cache-003191d3-fe38-4bde-8d7c-75b4aad413b4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.034868] env[62460]: DEBUG nova.compute.manager [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Instance network_info: |[{"id": "73df6556-a3c8-4ee6-9881-8b978f97ca6f", "address": "fa:16:3e:2b:5b:2e", "network": {"id": "0c9ffa5f-8a93-402d-bb98-8d88ef381c15", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-919853047-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28e73bd27f474a81a0aa7d4bb2d8efb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "450939f7-f74b-41f7-93f7-b4fde6a6fbed", "external-id": "nsx-vlan-transportzone-866", "segmentation_id": 866, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap73df6556-a3", "ovs_interfaceid": "73df6556-a3c8-4ee6-9881-8b978f97ca6f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 938.035267] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2b:5b:2e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '450939f7-f74b-41f7-93f7-b4fde6a6fbed', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '73df6556-a3c8-4ee6-9881-8b978f97ca6f', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 938.043413] env[62460]: DEBUG oslo.service.loopingcall [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 938.044655] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 938.044894] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-66e55560-708a-4993-8a6f-4be2e19287f3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.064153] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 938.064153] env[62460]: value = "task-1313892" [ 938.064153] env[62460]: _type = "Task" [ 938.064153] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.071905] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313892, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.119982] env[62460]: DEBUG nova.policy [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd7b38fb0a53f4d078ae86d68e3eb6268', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '53fa06008e9f43488362895e7a143700', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 938.158857] env[62460]: INFO nova.compute.manager [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] Took 15.11 seconds to build instance. [ 938.190714] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313890, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080589} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.191583] env[62460]: DEBUG nova.network.neutron [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Successfully updated port: 96b72a46-8d24-420d-a0f2-407651530be5 {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 938.194323] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 938.196077] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c60068c3-ca81-406f-a92c-d46594b855b5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.223540] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Reconfiguring VM instance instance-0000005a to attach disk [datastore1] 9c63946c-f294-42ea-979e-68db26d88858/9c63946c-f294-42ea-979e-68db26d88858.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 938.224269] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-64e8dda2-a7f6-471a-90ea-8fb851e08266 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.246384] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Waiting for the task: (returnval){ [ 938.246384] env[62460]: value = "task-1313893" [ 938.246384] env[62460]: _type = "Task" [ 938.246384] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.256288] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313893, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.402153] env[62460]: DEBUG oslo_vmware.api [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313891, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.522841] env[62460]: DEBUG nova.network.neutron [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Successfully created port: 90a7d774-4d76-4525-8e6d-8c08fdb93fb4 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 938.532290] env[62460]: DEBUG nova.compute.manager [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 938.536220] env[62460]: INFO nova.compute.claims [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 938.575196] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313892, 'name': CreateVM_Task, 'duration_secs': 0.507474} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.575196] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 938.575469] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.575630] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.576076] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 938.576385] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-293390f1-0763-451e-b135-0adfd89a8361 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.581771] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Waiting for the task: (returnval){ [ 938.581771] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52e21d95-4a8d-ad69-bb4e-610927d2b361" [ 938.581771] env[62460]: _type = "Task" [ 938.581771] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.590341] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52e21d95-4a8d-ad69-bb4e-610927d2b361, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.661357] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3d9c6820-dd02-435c-868d-2df9211c9120 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "6a5934ed-30dc-4512-8cf2-1e4d9f20836c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.621s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.688812] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8723b92f-e9a4-4205-a60c-fd54fe477845 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.695959] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Acquiring lock "refresh_cache-db331001-1744-4c42-a40c-87f3e60da4b8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.696121] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Acquired lock "refresh_cache-db331001-1744-4c42-a40c-87f3e60da4b8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.696295] env[62460]: DEBUG nova.network.neutron [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 938.697830] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-70448d4e-c376-4152-b16e-051a686ee848 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] Suspending the VM {{(pid=62460) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 938.698308] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-85a5d3e7-d7a4-40d2-ab4e-072b1ffd2dfa {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.705279] env[62460]: DEBUG oslo_vmware.api [None req-70448d4e-c376-4152-b16e-051a686ee848 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 938.705279] env[62460]: value = "task-1313894" [ 938.705279] env[62460]: _type = "Task" [ 938.705279] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.713201] env[62460]: DEBUG oslo_vmware.api [None req-70448d4e-c376-4152-b16e-051a686ee848 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313894, 'name': SuspendVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.756199] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313893, 'name': ReconfigVM_Task, 'duration_secs': 0.415099} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.756509] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Reconfigured VM instance instance-0000005a to attach disk [datastore1] 9c63946c-f294-42ea-979e-68db26d88858/9c63946c-f294-42ea-979e-68db26d88858.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 938.757155] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-35c8875b-2985-48a2-bfc9-cc9fc81da26b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.763439] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Waiting for the task: (returnval){ [ 938.763439] env[62460]: value = "task-1313895" [ 938.763439] env[62460]: _type = "Task" [ 938.763439] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.776461] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313895, 'name': Rename_Task} progress is 6%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.901236] env[62460]: DEBUG oslo_vmware.api [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313891, 'name': PowerOnVM_Task, 'duration_secs': 0.665689} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.901639] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 938.901729] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-939f7c00-7a1a-41f9-a672-709696784505 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Updating instance '6b084ed4-e8f0-4063-bf2a-252740640753' progress to 100 {{(pid=62460) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 939.092770] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52e21d95-4a8d-ad69-bb4e-610927d2b361, 'name': SearchDatastore_Task, 'duration_secs': 0.00877} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.093099] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.093346] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 939.093594] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.093899] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.093943] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 939.094205] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-af598836-34f6-4d59-b32d-1756e022d317 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.102481] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 939.102714] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 939.103443] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-01e50e75-f09d-4b15-9107-98792b3a3900 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.108299] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Waiting for the task: (returnval){ [ 939.108299] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52d588f6-1879-ab2c-b9d7-d36a99e48694" [ 939.108299] env[62460]: _type = "Task" [ 939.108299] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.115850] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52d588f6-1879-ab2c-b9d7-d36a99e48694, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.217184] env[62460]: DEBUG oslo_vmware.api [None req-70448d4e-c376-4152-b16e-051a686ee848 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313894, 'name': SuspendVM_Task} progress is 66%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.239523] env[62460]: DEBUG nova.network.neutron [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 939.276482] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313895, 'name': Rename_Task, 'duration_secs': 0.219053} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.276808] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 939.277103] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7ee099ab-e588-440d-aa2a-bf162ae982a8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.288826] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Waiting for the task: (returnval){ [ 939.288826] env[62460]: value = "task-1313896" [ 939.288826] env[62460]: _type = "Task" [ 939.288826] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.297325] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313896, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.422434] env[62460]: DEBUG nova.network.neutron [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Updating instance_info_cache with network_info: [{"id": "96b72a46-8d24-420d-a0f2-407651530be5", "address": "fa:16:3e:50:e8:9a", "network": {"id": "0c9ffa5f-8a93-402d-bb98-8d88ef381c15", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-919853047-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28e73bd27f474a81a0aa7d4bb2d8efb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "450939f7-f74b-41f7-93f7-b4fde6a6fbed", "external-id": "nsx-vlan-transportzone-866", "segmentation_id": 866, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap96b72a46-8d", "ovs_interfaceid": "96b72a46-8d24-420d-a0f2-407651530be5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 939.548214] env[62460]: DEBUG nova.compute.manager [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 939.574808] env[62460]: DEBUG nova.virt.hardware [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 939.575091] env[62460]: DEBUG nova.virt.hardware [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 939.575288] env[62460]: DEBUG nova.virt.hardware [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 939.575520] env[62460]: DEBUG nova.virt.hardware [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 939.575686] env[62460]: DEBUG nova.virt.hardware [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 939.575845] env[62460]: DEBUG nova.virt.hardware [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 939.576074] env[62460]: DEBUG nova.virt.hardware [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 939.576252] env[62460]: DEBUG nova.virt.hardware [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 939.576429] env[62460]: DEBUG nova.virt.hardware [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 939.576620] env[62460]: DEBUG nova.virt.hardware [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 939.576839] env[62460]: DEBUG nova.virt.hardware [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 939.577829] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ee20a15-d5c1-41a7-ae07-1f9c58b78f14 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.587078] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ff006c8-9c41-4775-bfdf-1ef8312f5adf {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.619385] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52d588f6-1879-ab2c-b9d7-d36a99e48694, 'name': SearchDatastore_Task, 'duration_secs': 0.009337} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.619940] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-99f629b5-4c02-4804-b90b-16dc70a4a4e1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.626633] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Waiting for the task: (returnval){ [ 939.626633] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]526ea7a3-006c-3d06-33f8-17fd058f5943" [ 939.626633] env[62460]: _type = "Task" [ 939.626633] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.634262] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]526ea7a3-006c-3d06-33f8-17fd058f5943, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.717524] env[62460]: DEBUG oslo_vmware.api [None req-70448d4e-c376-4152-b16e-051a686ee848 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313894, 'name': SuspendVM_Task, 'duration_secs': 0.611989} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.718904] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-70448d4e-c376-4152-b16e-051a686ee848 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] Suspended the VM {{(pid=62460) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 939.718904] env[62460]: DEBUG nova.compute.manager [None req-70448d4e-c376-4152-b16e-051a686ee848 tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 939.719494] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66d350ac-d15c-4a0a-aea6-dc5fa75cce83 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.722258] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0f9e9d3-988a-453a-b090-2377fe316be9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.730022] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf55e15e-17c1-4011-a8bb-d3f9a5133f5d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.762200] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8782b7fb-6be7-42a1-91ab-597f3912bbe6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.770060] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df782252-08cc-4f99-9d5e-1dadc3f16441 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.783721] env[62460]: DEBUG nova.compute.provider_tree [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 939.797944] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313896, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.926830] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Releasing lock "refresh_cache-db331001-1744-4c42-a40c-87f3e60da4b8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.927187] env[62460]: DEBUG nova.compute.manager [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Instance network_info: |[{"id": "96b72a46-8d24-420d-a0f2-407651530be5", "address": "fa:16:3e:50:e8:9a", "network": {"id": "0c9ffa5f-8a93-402d-bb98-8d88ef381c15", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-919853047-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28e73bd27f474a81a0aa7d4bb2d8efb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "450939f7-f74b-41f7-93f7-b4fde6a6fbed", "external-id": "nsx-vlan-transportzone-866", "segmentation_id": 866, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap96b72a46-8d", "ovs_interfaceid": "96b72a46-8d24-420d-a0f2-407651530be5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 939.927563] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:50:e8:9a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '450939f7-f74b-41f7-93f7-b4fde6a6fbed', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '96b72a46-8d24-420d-a0f2-407651530be5', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 939.936298] env[62460]: DEBUG oslo.service.loopingcall [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 939.936298] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 939.936298] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-38aa2d15-7ff4-411b-ba23-be5853efdca6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.956029] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 939.956029] env[62460]: value = "task-1313897" [ 939.956029] env[62460]: _type = "Task" [ 939.956029] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.963935] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313897, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.994140] env[62460]: DEBUG nova.compute.manager [req-1355de34-9a3c-49c4-8dd7-e26df907a56c req-06049647-ad1e-488f-be91-3173aa5daa63 service nova] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Received event network-changed-96b72a46-8d24-420d-a0f2-407651530be5 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 939.995708] env[62460]: DEBUG nova.compute.manager [req-1355de34-9a3c-49c4-8dd7-e26df907a56c req-06049647-ad1e-488f-be91-3173aa5daa63 service nova] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Refreshing instance network info cache due to event network-changed-96b72a46-8d24-420d-a0f2-407651530be5. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 939.995985] env[62460]: DEBUG oslo_concurrency.lockutils [req-1355de34-9a3c-49c4-8dd7-e26df907a56c req-06049647-ad1e-488f-be91-3173aa5daa63 service nova] Acquiring lock "refresh_cache-db331001-1744-4c42-a40c-87f3e60da4b8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.996209] env[62460]: DEBUG oslo_concurrency.lockutils [req-1355de34-9a3c-49c4-8dd7-e26df907a56c req-06049647-ad1e-488f-be91-3173aa5daa63 service nova] Acquired lock "refresh_cache-db331001-1744-4c42-a40c-87f3e60da4b8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.996418] env[62460]: DEBUG nova.network.neutron [req-1355de34-9a3c-49c4-8dd7-e26df907a56c req-06049647-ad1e-488f-be91-3173aa5daa63 service nova] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Refreshing network info cache for port 96b72a46-8d24-420d-a0f2-407651530be5 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 940.019429] env[62460]: DEBUG nova.network.neutron [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Successfully updated port: 90a7d774-4d76-4525-8e6d-8c08fdb93fb4 {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 940.138769] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]526ea7a3-006c-3d06-33f8-17fd058f5943, 'name': SearchDatastore_Task, 'duration_secs': 0.034785} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.139260] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.139570] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 003191d3-fe38-4bde-8d7c-75b4aad413b4/003191d3-fe38-4bde-8d7c-75b4aad413b4.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 940.139872] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-57011a2a-b9c6-4819-9f36-8f7e21e88320 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.146536] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Waiting for the task: (returnval){ [ 940.146536] env[62460]: value = "task-1313898" [ 940.146536] env[62460]: _type = "Task" [ 940.146536] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.154896] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313898, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.298413] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313896, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.315958] env[62460]: DEBUG nova.scheduler.client.report [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Updated inventory for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with generation 112 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 940.316268] env[62460]: DEBUG nova.compute.provider_tree [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Updating resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 generation from 112 to 113 during operation: update_inventory {{(pid=62460) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 940.316473] env[62460]: DEBUG nova.compute.provider_tree [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 940.466748] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313897, 'name': CreateVM_Task, 'duration_secs': 0.323958} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.466748] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 940.467512] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.467512] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.467865] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 940.468156] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-63faff56-a825-4ef3-93b3-ec2f4f5c8045 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.474086] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Waiting for the task: (returnval){ [ 940.474086] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52cd5107-1716-209d-f73b-46ed6fdde47f" [ 940.474086] env[62460]: _type = "Task" [ 940.474086] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.482455] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52cd5107-1716-209d-f73b-46ed6fdde47f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.522187] env[62460]: DEBUG oslo_concurrency.lockutils [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "refresh_cache-0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.522187] env[62460]: DEBUG oslo_concurrency.lockutils [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquired lock "refresh_cache-0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.522464] env[62460]: DEBUG nova.network.neutron [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 940.662867] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313898, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.801180] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313896, 'name': PowerOnVM_Task, 'duration_secs': 1.234238} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.801728] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 940.801728] env[62460]: INFO nova.compute.manager [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Took 9.19 seconds to spawn the instance on the hypervisor. [ 940.801972] env[62460]: DEBUG nova.compute.manager [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 940.803109] env[62460]: DEBUG nova.network.neutron [req-1355de34-9a3c-49c4-8dd7-e26df907a56c req-06049647-ad1e-488f-be91-3173aa5daa63 service nova] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Updated VIF entry in instance network info cache for port 96b72a46-8d24-420d-a0f2-407651530be5. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 940.803501] env[62460]: DEBUG nova.network.neutron [req-1355de34-9a3c-49c4-8dd7-e26df907a56c req-06049647-ad1e-488f-be91-3173aa5daa63 service nova] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Updating instance_info_cache with network_info: [{"id": "96b72a46-8d24-420d-a0f2-407651530be5", "address": "fa:16:3e:50:e8:9a", "network": {"id": "0c9ffa5f-8a93-402d-bb98-8d88ef381c15", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-919853047-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28e73bd27f474a81a0aa7d4bb2d8efb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "450939f7-f74b-41f7-93f7-b4fde6a6fbed", "external-id": "nsx-vlan-transportzone-866", "segmentation_id": 866, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap96b72a46-8d", "ovs_interfaceid": "96b72a46-8d24-420d-a0f2-407651530be5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.805523] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c4ab418-009b-404f-94fd-50557e4af3b8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.821604] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.303s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.824900] env[62460]: DEBUG oslo_concurrency.lockutils [None req-800eb51c-26fc-48d3-9ea6-3689387e9e69 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.811s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.824900] env[62460]: DEBUG nova.objects.instance [None req-800eb51c-26fc-48d3-9ea6-3689387e9e69 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lazy-loading 'resources' on Instance uuid 61073a4e-c5e5-48f5-a43a-ce71bd18e48a {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 940.864550] env[62460]: INFO nova.network.neutron [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Updating port 72a87ddf-0585-429a-b9de-d73bcad42cd1 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 940.985739] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52cd5107-1716-209d-f73b-46ed6fdde47f, 'name': SearchDatastore_Task, 'duration_secs': 0.009647} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.986098] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.986363] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 940.986618] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.986778] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.986966] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 940.987249] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d1a0f9c5-8413-4a92-bfbe-a9ef775aec14 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.995180] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 940.995389] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 940.996105] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-336fa8b2-75c8-4418-9bb5-29fc97811735 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.001610] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Waiting for the task: (returnval){ [ 941.001610] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]5249bbcb-227a-ff68-2590-abd5abb9af1e" [ 941.001610] env[62460]: _type = "Task" [ 941.001610] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.008679] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5249bbcb-227a-ff68-2590-abd5abb9af1e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.085046] env[62460]: DEBUG nova.network.neutron [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 941.161279] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313898, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.688515} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.161580] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 003191d3-fe38-4bde-8d7c-75b4aad413b4/003191d3-fe38-4bde-8d7c-75b4aad413b4.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 941.161792] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 941.162199] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3b909b9d-e5fc-4833-8455-ea6312842a24 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.170025] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Waiting for the task: (returnval){ [ 941.170025] env[62460]: value = "task-1313899" [ 941.170025] env[62460]: _type = "Task" [ 941.170025] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.182305] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313899, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.237829] env[62460]: DEBUG oslo_concurrency.lockutils [None req-35317be6-b47b-4771-852f-43d504ddaa7e tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "6a5934ed-30dc-4512-8cf2-1e4d9f20836c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.238138] env[62460]: DEBUG oslo_concurrency.lockutils [None req-35317be6-b47b-4771-852f-43d504ddaa7e tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "6a5934ed-30dc-4512-8cf2-1e4d9f20836c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.238381] env[62460]: DEBUG oslo_concurrency.lockutils [None req-35317be6-b47b-4771-852f-43d504ddaa7e tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "6a5934ed-30dc-4512-8cf2-1e4d9f20836c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.238586] env[62460]: DEBUG oslo_concurrency.lockutils [None req-35317be6-b47b-4771-852f-43d504ddaa7e tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "6a5934ed-30dc-4512-8cf2-1e4d9f20836c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.238762] env[62460]: DEBUG oslo_concurrency.lockutils [None req-35317be6-b47b-4771-852f-43d504ddaa7e tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "6a5934ed-30dc-4512-8cf2-1e4d9f20836c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.240975] env[62460]: INFO nova.compute.manager [None req-35317be6-b47b-4771-852f-43d504ddaa7e tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] Terminating instance [ 941.242758] env[62460]: DEBUG nova.compute.manager [None req-35317be6-b47b-4771-852f-43d504ddaa7e tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 941.242994] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-35317be6-b47b-4771-852f-43d504ddaa7e tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 941.243818] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8503a68e-e0ba-4c20-a4b5-a8544645e68b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.251449] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-35317be6-b47b-4771-852f-43d504ddaa7e tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 941.251685] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6ad58b91-a074-4f91-b69e-c8d5dcc93ffb {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.308166] env[62460]: DEBUG nova.network.neutron [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Updating instance_info_cache with network_info: [{"id": "90a7d774-4d76-4525-8e6d-8c08fdb93fb4", "address": "fa:16:3e:5b:1c:b7", "network": {"id": "d3136e32-ad55-4b73-835c-8fa5f0480767", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1511414968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53fa06008e9f43488362895e7a143700", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "054fcd1e-638e-425a-a1de-78cb188ae026", "external-id": "nsx-vlan-transportzone-658", "segmentation_id": 658, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90a7d774-4d", "ovs_interfaceid": "90a7d774-4d76-4525-8e6d-8c08fdb93fb4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.309826] env[62460]: DEBUG oslo_concurrency.lockutils [req-1355de34-9a3c-49c4-8dd7-e26df907a56c req-06049647-ad1e-488f-be91-3173aa5daa63 service nova] Releasing lock "refresh_cache-db331001-1744-4c42-a40c-87f3e60da4b8" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.310112] env[62460]: DEBUG nova.compute.manager [req-1355de34-9a3c-49c4-8dd7-e26df907a56c req-06049647-ad1e-488f-be91-3173aa5daa63 service nova] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Received event network-vif-plugged-90a7d774-4d76-4525-8e6d-8c08fdb93fb4 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 941.310348] env[62460]: DEBUG oslo_concurrency.lockutils [req-1355de34-9a3c-49c4-8dd7-e26df907a56c req-06049647-ad1e-488f-be91-3173aa5daa63 service nova] Acquiring lock "0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.310564] env[62460]: DEBUG oslo_concurrency.lockutils [req-1355de34-9a3c-49c4-8dd7-e26df907a56c req-06049647-ad1e-488f-be91-3173aa5daa63 service nova] Lock "0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.310736] env[62460]: DEBUG oslo_concurrency.lockutils [req-1355de34-9a3c-49c4-8dd7-e26df907a56c req-06049647-ad1e-488f-be91-3173aa5daa63 service nova] Lock "0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.310910] env[62460]: DEBUG nova.compute.manager [req-1355de34-9a3c-49c4-8dd7-e26df907a56c req-06049647-ad1e-488f-be91-3173aa5daa63 service nova] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] No waiting events found dispatching network-vif-plugged-90a7d774-4d76-4525-8e6d-8c08fdb93fb4 {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 941.311101] env[62460]: WARNING nova.compute.manager [req-1355de34-9a3c-49c4-8dd7-e26df907a56c req-06049647-ad1e-488f-be91-3173aa5daa63 service nova] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Received unexpected event network-vif-plugged-90a7d774-4d76-4525-8e6d-8c08fdb93fb4 for instance with vm_state building and task_state spawning. [ 941.312836] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-35317be6-b47b-4771-852f-43d504ddaa7e tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 941.313085] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-35317be6-b47b-4771-852f-43d504ddaa7e tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] Deleting contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 941.313223] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-35317be6-b47b-4771-852f-43d504ddaa7e tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Deleting the datastore file [datastore1] 6a5934ed-30dc-4512-8cf2-1e4d9f20836c {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 941.313512] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6b00b712-b199-4f36-a7f0-0a7988adfa6f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.325274] env[62460]: DEBUG oslo_vmware.api [None req-35317be6-b47b-4771-852f-43d504ddaa7e tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 941.325274] env[62460]: value = "task-1313901" [ 941.325274] env[62460]: _type = "Task" [ 941.325274] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.330434] env[62460]: INFO nova.compute.manager [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Took 17.08 seconds to build instance. [ 941.338994] env[62460]: DEBUG oslo_vmware.api [None req-35317be6-b47b-4771-852f-43d504ddaa7e tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313901, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.411495] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0f0685de-d89c-4234-9b4d-c36872e0a4f9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "6b084ed4-e8f0-4063-bf2a-252740640753" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.411784] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0f0685de-d89c-4234-9b4d-c36872e0a4f9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "6b084ed4-e8f0-4063-bf2a-252740640753" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.411982] env[62460]: DEBUG nova.compute.manager [None req-0f0685de-d89c-4234-9b4d-c36872e0a4f9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Going to confirm migration 1 {{(pid=62460) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 941.503465] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b18847c-3896-4ed7-9949-3e49f5a953f3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.516888] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-757f1da2-c965-4911-b9c9-fbeb8c6e2940 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.520237] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5249bbcb-227a-ff68-2590-abd5abb9af1e, 'name': SearchDatastore_Task, 'duration_secs': 0.01123} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.521240] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c8f65e23-8d15-4256-8e27-478856b072b7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.547227] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cf8fda7-b9fa-437b-b13d-22b831ea2d09 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.550662] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Waiting for the task: (returnval){ [ 941.550662] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]522ea179-4484-5f6d-92eb-bc6fdc3e18ca" [ 941.550662] env[62460]: _type = "Task" [ 941.550662] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.558917] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0db9e40-6ad2-439c-8d5d-39074367d738 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.565088] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]522ea179-4484-5f6d-92eb-bc6fdc3e18ca, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.574174] env[62460]: DEBUG nova.compute.provider_tree [None req-800eb51c-26fc-48d3-9ea6-3689387e9e69 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 941.679590] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313899, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.0638} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.680699] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 941.680699] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bc5bb91-6b89-452a-97df-c2942bd459f4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.703196] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Reconfiguring VM instance instance-0000005b to attach disk [datastore1] 003191d3-fe38-4bde-8d7c-75b4aad413b4/003191d3-fe38-4bde-8d7c-75b4aad413b4.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 941.703518] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-47732295-6947-43c6-9da1-d88a99f58e9d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.722812] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Waiting for the task: (returnval){ [ 941.722812] env[62460]: value = "task-1313902" [ 941.722812] env[62460]: _type = "Task" [ 941.722812] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.730424] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313902, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.812297] env[62460]: DEBUG oslo_concurrency.lockutils [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Releasing lock "refresh_cache-0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.813041] env[62460]: DEBUG nova.compute.manager [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Instance network_info: |[{"id": "90a7d774-4d76-4525-8e6d-8c08fdb93fb4", "address": "fa:16:3e:5b:1c:b7", "network": {"id": "d3136e32-ad55-4b73-835c-8fa5f0480767", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1511414968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53fa06008e9f43488362895e7a143700", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "054fcd1e-638e-425a-a1de-78cb188ae026", "external-id": "nsx-vlan-transportzone-658", "segmentation_id": 658, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90a7d774-4d", "ovs_interfaceid": "90a7d774-4d76-4525-8e6d-8c08fdb93fb4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 941.813041] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5b:1c:b7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '054fcd1e-638e-425a-a1de-78cb188ae026', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '90a7d774-4d76-4525-8e6d-8c08fdb93fb4', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 941.820817] env[62460]: DEBUG oslo.service.loopingcall [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 941.821079] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 941.821320] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7467b9d8-2a28-4310-817a-dbf0d04013e6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.835503] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Lock "9c63946c-f294-42ea-979e-68db26d88858" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.594s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.844800] env[62460]: DEBUG oslo_vmware.api [None req-35317be6-b47b-4771-852f-43d504ddaa7e tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313901, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.487959} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.845050] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-35317be6-b47b-4771-852f-43d504ddaa7e tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 941.845243] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-35317be6-b47b-4771-852f-43d504ddaa7e tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] Deleted contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 941.845426] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-35317be6-b47b-4771-852f-43d504ddaa7e tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 941.845607] env[62460]: INFO nova.compute.manager [None req-35317be6-b47b-4771-852f-43d504ddaa7e tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] Took 0.60 seconds to destroy the instance on the hypervisor. [ 941.845842] env[62460]: DEBUG oslo.service.loopingcall [None req-35317be6-b47b-4771-852f-43d504ddaa7e tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 941.846044] env[62460]: DEBUG nova.compute.manager [-] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 941.846144] env[62460]: DEBUG nova.network.neutron [-] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 941.848896] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 941.848896] env[62460]: value = "task-1313903" [ 941.848896] env[62460]: _type = "Task" [ 941.848896] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.857691] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313903, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.991918] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0f0685de-d89c-4234-9b4d-c36872e0a4f9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "refresh_cache-6b084ed4-e8f0-4063-bf2a-252740640753" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.992307] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0f0685de-d89c-4234-9b4d-c36872e0a4f9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquired lock "refresh_cache-6b084ed4-e8f0-4063-bf2a-252740640753" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.992307] env[62460]: DEBUG nova.network.neutron [None req-0f0685de-d89c-4234-9b4d-c36872e0a4f9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 941.992561] env[62460]: DEBUG nova.objects.instance [None req-0f0685de-d89c-4234-9b4d-c36872e0a4f9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lazy-loading 'info_cache' on Instance uuid 6b084ed4-e8f0-4063-bf2a-252740640753 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 942.022838] env[62460]: DEBUG nova.compute.manager [req-111f898f-51c7-4986-87d9-1e35921312b1 req-05b483e9-e1af-4747-9d07-c2214d8c47c3 service nova] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Received event network-changed-90a7d774-4d76-4525-8e6d-8c08fdb93fb4 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 942.022970] env[62460]: DEBUG nova.compute.manager [req-111f898f-51c7-4986-87d9-1e35921312b1 req-05b483e9-e1af-4747-9d07-c2214d8c47c3 service nova] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Refreshing instance network info cache due to event network-changed-90a7d774-4d76-4525-8e6d-8c08fdb93fb4. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 942.023202] env[62460]: DEBUG oslo_concurrency.lockutils [req-111f898f-51c7-4986-87d9-1e35921312b1 req-05b483e9-e1af-4747-9d07-c2214d8c47c3 service nova] Acquiring lock "refresh_cache-0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.023368] env[62460]: DEBUG oslo_concurrency.lockutils [req-111f898f-51c7-4986-87d9-1e35921312b1 req-05b483e9-e1af-4747-9d07-c2214d8c47c3 service nova] Acquired lock "refresh_cache-0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.023540] env[62460]: DEBUG nova.network.neutron [req-111f898f-51c7-4986-87d9-1e35921312b1 req-05b483e9-e1af-4747-9d07-c2214d8c47c3 service nova] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Refreshing network info cache for port 90a7d774-4d76-4525-8e6d-8c08fdb93fb4 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 942.062615] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]522ea179-4484-5f6d-92eb-bc6fdc3e18ca, 'name': SearchDatastore_Task, 'duration_secs': 0.037244} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.062873] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 942.063166] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] db331001-1744-4c42-a40c-87f3e60da4b8/db331001-1744-4c42-a40c-87f3e60da4b8.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 942.063702] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3ded6317-a2c1-4e83-8467-5bafba539acc {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.070648] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Waiting for the task: (returnval){ [ 942.070648] env[62460]: value = "task-1313904" [ 942.070648] env[62460]: _type = "Task" [ 942.070648] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.080713] env[62460]: DEBUG nova.scheduler.client.report [None req-800eb51c-26fc-48d3-9ea6-3689387e9e69 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 942.084788] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313904, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.233489] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313902, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.359270] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313903, 'name': CreateVM_Task} progress is 25%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.530684] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquiring lock "refresh_cache-896f8ff4-e45f-4403-a727-03ee25e58609" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.530948] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquired lock "refresh_cache-896f8ff4-e45f-4403-a727-03ee25e58609" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.531088] env[62460]: DEBUG nova.network.neutron [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 942.581799] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313904, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.585634] env[62460]: DEBUG oslo_concurrency.lockutils [None req-800eb51c-26fc-48d3-9ea6-3689387e9e69 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.761s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.588058] env[62460]: DEBUG oslo_concurrency.lockutils [None req-18074529-fe3b-4b5f-9546-b5cf5e11a600 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.453s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.588339] env[62460]: DEBUG nova.objects.instance [None req-18074529-fe3b-4b5f-9546-b5cf5e11a600 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Lazy-loading 'resources' on Instance uuid 3e10af70-db52-4d4a-bb92-821a05dcbab6 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 942.589407] env[62460]: DEBUG nova.network.neutron [-] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 942.632063] env[62460]: INFO nova.scheduler.client.report [None req-800eb51c-26fc-48d3-9ea6-3689387e9e69 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Deleted allocations for instance 61073a4e-c5e5-48f5-a43a-ce71bd18e48a [ 942.740494] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313902, 'name': ReconfigVM_Task, 'duration_secs': 0.885324} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.740838] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Reconfigured VM instance instance-0000005b to attach disk [datastore1] 003191d3-fe38-4bde-8d7c-75b4aad413b4/003191d3-fe38-4bde-8d7c-75b4aad413b4.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 942.742733] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2236d85e-5ca2-4c03-b533-79f0c34d15aa {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.756212] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Waiting for the task: (returnval){ [ 942.756212] env[62460]: value = "task-1313905" [ 942.756212] env[62460]: _type = "Task" [ 942.756212] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.761765] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313905, 'name': Rename_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.771279] env[62460]: DEBUG nova.network.neutron [req-111f898f-51c7-4986-87d9-1e35921312b1 req-05b483e9-e1af-4747-9d07-c2214d8c47c3 service nova] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Updated VIF entry in instance network info cache for port 90a7d774-4d76-4525-8e6d-8c08fdb93fb4. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 942.771645] env[62460]: DEBUG nova.network.neutron [req-111f898f-51c7-4986-87d9-1e35921312b1 req-05b483e9-e1af-4747-9d07-c2214d8c47c3 service nova] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Updating instance_info_cache with network_info: [{"id": "90a7d774-4d76-4525-8e6d-8c08fdb93fb4", "address": "fa:16:3e:5b:1c:b7", "network": {"id": "d3136e32-ad55-4b73-835c-8fa5f0480767", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1511414968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53fa06008e9f43488362895e7a143700", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "054fcd1e-638e-425a-a1de-78cb188ae026", "external-id": "nsx-vlan-transportzone-658", "segmentation_id": 658, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90a7d774-4d", "ovs_interfaceid": "90a7d774-4d76-4525-8e6d-8c08fdb93fb4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 942.858316] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313903, 'name': CreateVM_Task, 'duration_secs': 0.716901} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.858515] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 942.859487] env[62460]: DEBUG oslo_concurrency.lockutils [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.859487] env[62460]: DEBUG oslo_concurrency.lockutils [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.859739] env[62460]: DEBUG oslo_concurrency.lockutils [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 942.859998] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7fec0d46-62c8-4261-bf1d-b5e780bab398 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.864425] env[62460]: DEBUG oslo_vmware.api [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Waiting for the task: (returnval){ [ 942.864425] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]527c535a-d66f-65fe-5424-fc5a494d9e84" [ 942.864425] env[62460]: _type = "Task" [ 942.864425] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.872336] env[62460]: DEBUG oslo_vmware.api [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]527c535a-d66f-65fe-5424-fc5a494d9e84, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.083979] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313904, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.766506} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.084304] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] db331001-1744-4c42-a40c-87f3e60da4b8/db331001-1744-4c42-a40c-87f3e60da4b8.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 943.084544] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 943.084844] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2b7dac05-8211-4698-a040-74f2ccd49403 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.098768] env[62460]: INFO nova.compute.manager [-] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] Took 1.25 seconds to deallocate network for instance. [ 943.100602] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Waiting for the task: (returnval){ [ 943.100602] env[62460]: value = "task-1313906" [ 943.100602] env[62460]: _type = "Task" [ 943.100602] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.112668] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313906, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.140172] env[62460]: DEBUG oslo_concurrency.lockutils [None req-800eb51c-26fc-48d3-9ea6-3689387e9e69 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "61073a4e-c5e5-48f5-a43a-ce71bd18e48a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.057s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.265247] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313905, 'name': Rename_Task, 'duration_secs': 0.178754} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.265554] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 943.265809] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-30573be6-9ae1-4f8a-91d5-5da9f22d862c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.272435] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Waiting for the task: (returnval){ [ 943.272435] env[62460]: value = "task-1313907" [ 943.272435] env[62460]: _type = "Task" [ 943.272435] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.276084] env[62460]: DEBUG oslo_concurrency.lockutils [req-111f898f-51c7-4986-87d9-1e35921312b1 req-05b483e9-e1af-4747-9d07-c2214d8c47c3 service nova] Releasing lock "refresh_cache-0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.282445] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313907, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.297971] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f31e64d7-f774-49f3-b5e8-0c06441963e1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.302452] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4fe60fd1-0c1e-413d-b86f-9f13b8083746 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "e90e8c3a-d244-4d04-997e-c15ab1cb2009" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.302757] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4fe60fd1-0c1e-413d-b86f-9f13b8083746 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "e90e8c3a-d244-4d04-997e-c15ab1cb2009" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.303010] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4fe60fd1-0c1e-413d-b86f-9f13b8083746 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "e90e8c3a-d244-4d04-997e-c15ab1cb2009-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.303227] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4fe60fd1-0c1e-413d-b86f-9f13b8083746 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "e90e8c3a-d244-4d04-997e-c15ab1cb2009-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.303405] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4fe60fd1-0c1e-413d-b86f-9f13b8083746 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "e90e8c3a-d244-4d04-997e-c15ab1cb2009-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.307561] env[62460]: INFO nova.compute.manager [None req-4fe60fd1-0c1e-413d-b86f-9f13b8083746 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] Terminating instance [ 943.309337] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59995bcf-680c-4c43-9730-4056ac096111 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.313216] env[62460]: DEBUG nova.compute.manager [None req-4fe60fd1-0c1e-413d-b86f-9f13b8083746 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 943.313728] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-4fe60fd1-0c1e-413d-b86f-9f13b8083746 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 943.314145] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72b03428-2eb8-449b-b70a-f2bd0f9436dc {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.345681] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fe60fd1-0c1e-413d-b86f-9f13b8083746 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 943.345815] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dfdae194-1f3d-4a93-b711-37f780af3221 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.350043] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9d03526-09b8-453a-8197-71ae2caf8ccc {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.356729] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8039748-1a85-48d0-a407-c8a474f9a774 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.360973] env[62460]: DEBUG oslo_vmware.api [None req-4fe60fd1-0c1e-413d-b86f-9f13b8083746 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for the task: (returnval){ [ 943.360973] env[62460]: value = "task-1313908" [ 943.360973] env[62460]: _type = "Task" [ 943.360973] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.374260] env[62460]: DEBUG nova.compute.provider_tree [None req-18074529-fe3b-4b5f-9546-b5cf5e11a600 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 943.376386] env[62460]: DEBUG nova.network.neutron [None req-0f0685de-d89c-4234-9b4d-c36872e0a4f9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Updating instance_info_cache with network_info: [{"id": "3eaf04a4-feda-4a00-908e-d121fc775ecb", "address": "fa:16:3e:98:27:28", "network": {"id": "3c49720c-94c9-4e2f-bbb7-4266cc97faee", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1650281790-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfa5a65b0f614e769de5b3aa77bf869e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3eaf04a4-fe", "ovs_interfaceid": "3eaf04a4-feda-4a00-908e-d121fc775ecb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.385894] env[62460]: DEBUG oslo_vmware.api [None req-4fe60fd1-0c1e-413d-b86f-9f13b8083746 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313908, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.392376] env[62460]: DEBUG oslo_vmware.api [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]527c535a-d66f-65fe-5424-fc5a494d9e84, 'name': SearchDatastore_Task, 'duration_secs': 0.008214} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.392674] env[62460]: DEBUG oslo_concurrency.lockutils [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.392896] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 943.393482] env[62460]: DEBUG oslo_concurrency.lockutils [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.393482] env[62460]: DEBUG oslo_concurrency.lockutils [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.394296] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 943.394296] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-90fbe649-a44d-49e9-99c5-f63dada3326e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.401719] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 943.403732] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 943.403732] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fd21251c-5704-415c-9814-c97804adb9c1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.407771] env[62460]: DEBUG oslo_vmware.api [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Waiting for the task: (returnval){ [ 943.407771] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52c0bf96-a750-5947-716e-7f12aeee6f58" [ 943.407771] env[62460]: _type = "Task" [ 943.407771] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.416248] env[62460]: DEBUG oslo_vmware.api [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52c0bf96-a750-5947-716e-7f12aeee6f58, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.428727] env[62460]: DEBUG nova.network.neutron [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Updating instance_info_cache with network_info: [{"id": "72a87ddf-0585-429a-b9de-d73bcad42cd1", "address": "fa:16:3e:85:2f:50", "network": {"id": "b5cb4ee0-4562-45be-a25d-51c3a5932efd", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-823340443-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.149", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3d2d1c48ec14121a2e8c9b3f800a949", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55c757ac-f8b2-466d-b634-07dbd100b312", "external-id": "nsx-vlan-transportzone-159", "segmentation_id": 159, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap72a87ddf-05", "ovs_interfaceid": "72a87ddf-0585-429a-b9de-d73bcad42cd1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.611481] env[62460]: DEBUG oslo_concurrency.lockutils [None req-35317be6-b47b-4771-852f-43d504ddaa7e tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.611797] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313906, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065644} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.612061] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 943.612835] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d0a24b2-6fc4-44fe-a923-a50d1046d063 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.634944] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Reconfiguring VM instance instance-0000005c to attach disk [datastore1] db331001-1744-4c42-a40c-87f3e60da4b8/db331001-1744-4c42-a40c-87f3e60da4b8.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 943.635242] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-37a819db-b289-4c7a-b963-f7a369a69715 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.654507] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Waiting for the task: (returnval){ [ 943.654507] env[62460]: value = "task-1313909" [ 943.654507] env[62460]: _type = "Task" [ 943.654507] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.663805] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313909, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.781645] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313907, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.870676] env[62460]: DEBUG oslo_vmware.api [None req-4fe60fd1-0c1e-413d-b86f-9f13b8083746 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313908, 'name': PowerOffVM_Task, 'duration_secs': 0.390262} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.870845] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fe60fd1-0c1e-413d-b86f-9f13b8083746 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 943.871023] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-4fe60fd1-0c1e-413d-b86f-9f13b8083746 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 943.872792] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bceda427-cd2e-4963-8b4f-fc3b3c388183 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.884905] env[62460]: DEBUG nova.scheduler.client.report [None req-18074529-fe3b-4b5f-9546-b5cf5e11a600 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 943.888161] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0f0685de-d89c-4234-9b4d-c36872e0a4f9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Releasing lock "refresh_cache-6b084ed4-e8f0-4063-bf2a-252740640753" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.888392] env[62460]: DEBUG nova.objects.instance [None req-0f0685de-d89c-4234-9b4d-c36872e0a4f9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lazy-loading 'migration_context' on Instance uuid 6b084ed4-e8f0-4063-bf2a-252740640753 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 943.918388] env[62460]: DEBUG oslo_vmware.api [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52c0bf96-a750-5947-716e-7f12aeee6f58, 'name': SearchDatastore_Task, 'duration_secs': 0.010324} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.919222] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-83feccde-69e3-4214-8221-29b2a0563850 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.924691] env[62460]: DEBUG oslo_vmware.api [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Waiting for the task: (returnval){ [ 943.924691] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52d43601-3849-bb3b-0b3a-9b5664c0897f" [ 943.924691] env[62460]: _type = "Task" [ 943.924691] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.932704] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Releasing lock "refresh_cache-896f8ff4-e45f-4403-a727-03ee25e58609" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.934833] env[62460]: DEBUG oslo_vmware.api [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52d43601-3849-bb3b-0b3a-9b5664c0897f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.939278] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-4fe60fd1-0c1e-413d-b86f-9f13b8083746 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 943.939484] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-4fe60fd1-0c1e-413d-b86f-9f13b8083746 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] Deleting contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 943.939672] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-4fe60fd1-0c1e-413d-b86f-9f13b8083746 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Deleting the datastore file [datastore1] e90e8c3a-d244-4d04-997e-c15ab1cb2009 {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 943.939928] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ba625728-34b6-40d1-a58f-cad6b2b51cc9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.946152] env[62460]: DEBUG oslo_vmware.api [None req-4fe60fd1-0c1e-413d-b86f-9f13b8083746 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for the task: (returnval){ [ 943.946152] env[62460]: value = "task-1313911" [ 943.946152] env[62460]: _type = "Task" [ 943.946152] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.954135] env[62460]: DEBUG oslo_vmware.api [None req-4fe60fd1-0c1e-413d-b86f-9f13b8083746 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313911, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.961836] env[62460]: DEBUG nova.virt.hardware [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='e7ee83c5135744112721f7c0bd57a4f6',container_format='bare',created_at=2024-10-18T09:06:22Z,direct_url=,disk_format='vmdk',id=99119861-894e-4add-bc8a-5414d6c25b72,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-943543148-shelved',owner='f3d2d1c48ec14121a2e8c9b3f800a949',properties=ImageMetaProps,protected=,size=31669248,status='active',tags=,updated_at=2024-10-18T09:06:34Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 943.962162] env[62460]: DEBUG nova.virt.hardware [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 943.962348] env[62460]: DEBUG nova.virt.hardware [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 943.962556] env[62460]: DEBUG nova.virt.hardware [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 943.962742] env[62460]: DEBUG nova.virt.hardware [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 943.962903] env[62460]: DEBUG nova.virt.hardware [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 943.963195] env[62460]: DEBUG nova.virt.hardware [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 943.963482] env[62460]: DEBUG nova.virt.hardware [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 943.963703] env[62460]: DEBUG nova.virt.hardware [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 943.963884] env[62460]: DEBUG nova.virt.hardware [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 943.964090] env[62460]: DEBUG nova.virt.hardware [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 943.964877] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e69e77f-d192-4233-829c-a46283a67b15 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.972193] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b8c88a3-c5a6-4c70-a2a2-eeb55edceb45 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.986842] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:85:2f:50', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '55c757ac-f8b2-466d-b634-07dbd100b312', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '72a87ddf-0585-429a-b9de-d73bcad42cd1', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 943.994100] env[62460]: DEBUG oslo.service.loopingcall [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 943.994353] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 943.994567] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8b991e57-5312-49bb-a369-26f05eb0c9cc {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.012335] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 944.012335] env[62460]: value = "task-1313912" [ 944.012335] env[62460]: _type = "Task" [ 944.012335] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.020025] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313912, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.063220] env[62460]: DEBUG nova.compute.manager [req-7042657d-271b-40f7-ab23-01a4ca4e2eea req-e859bbdc-9d0c-4c86-87d8-c7d71e9ca457 service nova] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] Received event network-vif-deleted-555d2fbe-16ac-4dd3-baa7-5b3b04b75670 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 944.063439] env[62460]: DEBUG nova.compute.manager [req-7042657d-271b-40f7-ab23-01a4ca4e2eea req-e859bbdc-9d0c-4c86-87d8-c7d71e9ca457 service nova] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Received event network-vif-plugged-72a87ddf-0585-429a-b9de-d73bcad42cd1 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 944.063727] env[62460]: DEBUG oslo_concurrency.lockutils [req-7042657d-271b-40f7-ab23-01a4ca4e2eea req-e859bbdc-9d0c-4c86-87d8-c7d71e9ca457 service nova] Acquiring lock "896f8ff4-e45f-4403-a727-03ee25e58609-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.064021] env[62460]: DEBUG oslo_concurrency.lockutils [req-7042657d-271b-40f7-ab23-01a4ca4e2eea req-e859bbdc-9d0c-4c86-87d8-c7d71e9ca457 service nova] Lock "896f8ff4-e45f-4403-a727-03ee25e58609-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.064137] env[62460]: DEBUG oslo_concurrency.lockutils [req-7042657d-271b-40f7-ab23-01a4ca4e2eea req-e859bbdc-9d0c-4c86-87d8-c7d71e9ca457 service nova] Lock "896f8ff4-e45f-4403-a727-03ee25e58609-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.064309] env[62460]: DEBUG nova.compute.manager [req-7042657d-271b-40f7-ab23-01a4ca4e2eea req-e859bbdc-9d0c-4c86-87d8-c7d71e9ca457 service nova] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] No waiting events found dispatching network-vif-plugged-72a87ddf-0585-429a-b9de-d73bcad42cd1 {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 944.064506] env[62460]: WARNING nova.compute.manager [req-7042657d-271b-40f7-ab23-01a4ca4e2eea req-e859bbdc-9d0c-4c86-87d8-c7d71e9ca457 service nova] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Received unexpected event network-vif-plugged-72a87ddf-0585-429a-b9de-d73bcad42cd1 for instance with vm_state shelved_offloaded and task_state spawning. [ 944.064719] env[62460]: DEBUG nova.compute.manager [req-7042657d-271b-40f7-ab23-01a4ca4e2eea req-e859bbdc-9d0c-4c86-87d8-c7d71e9ca457 service nova] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Received event network-changed-72a87ddf-0585-429a-b9de-d73bcad42cd1 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 944.065749] env[62460]: DEBUG nova.compute.manager [req-7042657d-271b-40f7-ab23-01a4ca4e2eea req-e859bbdc-9d0c-4c86-87d8-c7d71e9ca457 service nova] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Refreshing instance network info cache due to event network-changed-72a87ddf-0585-429a-b9de-d73bcad42cd1. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 944.065749] env[62460]: DEBUG oslo_concurrency.lockutils [req-7042657d-271b-40f7-ab23-01a4ca4e2eea req-e859bbdc-9d0c-4c86-87d8-c7d71e9ca457 service nova] Acquiring lock "refresh_cache-896f8ff4-e45f-4403-a727-03ee25e58609" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 944.065749] env[62460]: DEBUG oslo_concurrency.lockutils [req-7042657d-271b-40f7-ab23-01a4ca4e2eea req-e859bbdc-9d0c-4c86-87d8-c7d71e9ca457 service nova] Acquired lock "refresh_cache-896f8ff4-e45f-4403-a727-03ee25e58609" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.065749] env[62460]: DEBUG nova.network.neutron [req-7042657d-271b-40f7-ab23-01a4ca4e2eea req-e859bbdc-9d0c-4c86-87d8-c7d71e9ca457 service nova] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Refreshing network info cache for port 72a87ddf-0585-429a-b9de-d73bcad42cd1 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 944.164987] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313909, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.282435] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313907, 'name': PowerOnVM_Task} progress is 87%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.390832] env[62460]: DEBUG oslo_concurrency.lockutils [None req-18074529-fe3b-4b5f-9546-b5cf5e11a600 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.803s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.394338] env[62460]: DEBUG nova.objects.base [None req-0f0685de-d89c-4234-9b4d-c36872e0a4f9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Object Instance<6b084ed4-e8f0-4063-bf2a-252740640753> lazy-loaded attributes: info_cache,migration_context {{(pid=62460) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 944.394338] env[62460]: DEBUG oslo_concurrency.lockutils [None req-35317be6-b47b-4771-852f-43d504ddaa7e tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.783s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.394778] env[62460]: DEBUG nova.objects.instance [None req-35317be6-b47b-4771-852f-43d504ddaa7e tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lazy-loading 'resources' on Instance uuid 6a5934ed-30dc-4512-8cf2-1e4d9f20836c {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 944.396026] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb9fdb9f-f7f2-4506-a7d3-58908e771c50 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.418652] env[62460]: INFO nova.scheduler.client.report [None req-18074529-fe3b-4b5f-9546-b5cf5e11a600 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Deleted allocations for instance 3e10af70-db52-4d4a-bb92-821a05dcbab6 [ 944.420259] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-637386cd-6549-4ba1-b1df-00b19fdf5b10 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.430346] env[62460]: DEBUG oslo_vmware.api [None req-0f0685de-d89c-4234-9b4d-c36872e0a4f9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 944.430346] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52a94194-ba80-2672-5624-1f4b34cb70b1" [ 944.430346] env[62460]: _type = "Task" [ 944.430346] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.437494] env[62460]: DEBUG oslo_vmware.api [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52d43601-3849-bb3b-0b3a-9b5664c0897f, 'name': SearchDatastore_Task, 'duration_secs': 0.017455} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.438119] env[62460]: DEBUG oslo_concurrency.lockutils [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.438397] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea/0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 944.438691] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-82340351-f328-4ffa-9bcc-544e24ad3055 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.443362] env[62460]: DEBUG oslo_vmware.api [None req-0f0685de-d89c-4234-9b4d-c36872e0a4f9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52a94194-ba80-2672-5624-1f4b34cb70b1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.447958] env[62460]: DEBUG oslo_vmware.api [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Waiting for the task: (returnval){ [ 944.447958] env[62460]: value = "task-1313913" [ 944.447958] env[62460]: _type = "Task" [ 944.447958] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.459144] env[62460]: DEBUG oslo_vmware.api [None req-4fe60fd1-0c1e-413d-b86f-9f13b8083746 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313911, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.380938} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.462263] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-4fe60fd1-0c1e-413d-b86f-9f13b8083746 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 944.462597] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-4fe60fd1-0c1e-413d-b86f-9f13b8083746 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] Deleted contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 944.462700] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-4fe60fd1-0c1e-413d-b86f-9f13b8083746 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 944.462830] env[62460]: INFO nova.compute.manager [None req-4fe60fd1-0c1e-413d-b86f-9f13b8083746 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] Took 1.15 seconds to destroy the instance on the hypervisor. [ 944.463093] env[62460]: DEBUG oslo.service.loopingcall [None req-4fe60fd1-0c1e-413d-b86f-9f13b8083746 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 944.463314] env[62460]: DEBUG oslo_vmware.api [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313913, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.463879] env[62460]: DEBUG nova.compute.manager [-] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 944.463978] env[62460]: DEBUG nova.network.neutron [-] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 944.521995] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313912, 'name': CreateVM_Task} progress is 25%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.667871] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313909, 'name': ReconfigVM_Task, 'duration_secs': 0.76682} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.668535] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Reconfigured VM instance instance-0000005c to attach disk [datastore1] db331001-1744-4c42-a40c-87f3e60da4b8/db331001-1744-4c42-a40c-87f3e60da4b8.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 944.669394] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6abc1a73-171d-4306-8b16-d22998eb272d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.676178] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Waiting for the task: (returnval){ [ 944.676178] env[62460]: value = "task-1313914" [ 944.676178] env[62460]: _type = "Task" [ 944.676178] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.687585] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313914, 'name': Rename_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.782153] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313907, 'name': PowerOnVM_Task, 'duration_secs': 1.259108} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.782466] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 944.782684] env[62460]: INFO nova.compute.manager [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Took 9.89 seconds to spawn the instance on the hypervisor. [ 944.782878] env[62460]: DEBUG nova.compute.manager [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 944.783699] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf246e27-38f4-44ac-8717-24e13a9ee062 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.925273] env[62460]: DEBUG nova.network.neutron [req-7042657d-271b-40f7-ab23-01a4ca4e2eea req-e859bbdc-9d0c-4c86-87d8-c7d71e9ca457 service nova] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Updated VIF entry in instance network info cache for port 72a87ddf-0585-429a-b9de-d73bcad42cd1. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 944.925730] env[62460]: DEBUG nova.network.neutron [req-7042657d-271b-40f7-ab23-01a4ca4e2eea req-e859bbdc-9d0c-4c86-87d8-c7d71e9ca457 service nova] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Updating instance_info_cache with network_info: [{"id": "72a87ddf-0585-429a-b9de-d73bcad42cd1", "address": "fa:16:3e:85:2f:50", "network": {"id": "b5cb4ee0-4562-45be-a25d-51c3a5932efd", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-823340443-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.149", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3d2d1c48ec14121a2e8c9b3f800a949", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55c757ac-f8b2-466d-b634-07dbd100b312", "external-id": "nsx-vlan-transportzone-159", "segmentation_id": 159, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap72a87ddf-05", "ovs_interfaceid": "72a87ddf-0585-429a-b9de-d73bcad42cd1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 944.932911] env[62460]: DEBUG oslo_concurrency.lockutils [None req-18074529-fe3b-4b5f-9546-b5cf5e11a600 tempest-AttachVolumeNegativeTest-401902439 tempest-AttachVolumeNegativeTest-401902439-project-member] Lock "3e10af70-db52-4d4a-bb92-821a05dcbab6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.897s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.946414] env[62460]: DEBUG oslo_vmware.api [None req-0f0685de-d89c-4234-9b4d-c36872e0a4f9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52a94194-ba80-2672-5624-1f4b34cb70b1, 'name': SearchDatastore_Task, 'duration_secs': 0.014261} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.949176] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0f0685de-d89c-4234-9b4d-c36872e0a4f9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.959636] env[62460]: DEBUG oslo_vmware.api [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313913, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.025376] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313912, 'name': CreateVM_Task, 'duration_secs': 0.639599} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.025599] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 945.026288] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/99119861-894e-4add-bc8a-5414d6c25b72" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.026488] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquired lock "[datastore2] devstack-image-cache_base/99119861-894e-4add-bc8a-5414d6c25b72" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.026856] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/99119861-894e-4add-bc8a-5414d6c25b72" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 945.029589] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-344204e6-a4ea-4e2f-80bb-90bba45fcbf9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.034194] env[62460]: DEBUG oslo_vmware.api [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 945.034194] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52747d8b-a4fd-c019-bde0-f176aa7f07e2" [ 945.034194] env[62460]: _type = "Task" [ 945.034194] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.042154] env[62460]: DEBUG oslo_vmware.api [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52747d8b-a4fd-c019-bde0-f176aa7f07e2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.076544] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cbdd507-9614-4b02-9d8f-cc90a231493e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.083710] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b01f10e-350a-4900-8a1c-490d4a3ce2d2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.114954] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-493f7421-3bab-4c7a-ab12-e7113e0fb0ba {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.123779] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08420f3f-5301-4a4c-83cb-8baf0f02db4c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.136028] env[62460]: DEBUG nova.compute.provider_tree [None req-35317be6-b47b-4771-852f-43d504ddaa7e tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 945.186238] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313914, 'name': Rename_Task, 'duration_secs': 0.263127} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.186782] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 945.186782] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e646ed60-4437-4cb6-9765-22bf67be91c2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.195106] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Waiting for the task: (returnval){ [ 945.195106] env[62460]: value = "task-1313915" [ 945.195106] env[62460]: _type = "Task" [ 945.195106] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.202681] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313915, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.300970] env[62460]: INFO nova.compute.manager [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Took 21.01 seconds to build instance. [ 945.374729] env[62460]: DEBUG nova.network.neutron [-] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 945.431370] env[62460]: DEBUG oslo_concurrency.lockutils [req-7042657d-271b-40f7-ab23-01a4ca4e2eea req-e859bbdc-9d0c-4c86-87d8-c7d71e9ca457 service nova] Releasing lock "refresh_cache-896f8ff4-e45f-4403-a727-03ee25e58609" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.462036] env[62460]: DEBUG oslo_vmware.api [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313913, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.524584} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.462036] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea/0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 945.462036] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 945.462410] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-38d2dc40-7d4d-447b-bb4a-cf9554deba88 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.469871] env[62460]: DEBUG oslo_vmware.api [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Waiting for the task: (returnval){ [ 945.469871] env[62460]: value = "task-1313917" [ 945.469871] env[62460]: _type = "Task" [ 945.469871] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.477400] env[62460]: DEBUG oslo_vmware.api [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313917, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.544620] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Releasing lock "[datastore2] devstack-image-cache_base/99119861-894e-4add-bc8a-5414d6c25b72" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.544892] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Processing image 99119861-894e-4add-bc8a-5414d6c25b72 {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 945.545182] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/99119861-894e-4add-bc8a-5414d6c25b72/99119861-894e-4add-bc8a-5414d6c25b72.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.545324] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquired lock "[datastore2] devstack-image-cache_base/99119861-894e-4add-bc8a-5414d6c25b72/99119861-894e-4add-bc8a-5414d6c25b72.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.545512] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 945.545772] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2fc64c22-b2ea-4043-925d-89758ac6ba82 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.555039] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 945.555296] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 945.556069] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-df2bf7bb-253f-4df4-9dea-29c98d8cbdd3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.561779] env[62460]: DEBUG oslo_vmware.api [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 945.561779] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]525d6b0a-4c89-8460-5f83-fbe43734f32b" [ 945.561779] env[62460]: _type = "Task" [ 945.561779] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.570406] env[62460]: DEBUG oslo_vmware.api [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]525d6b0a-4c89-8460-5f83-fbe43734f32b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.638863] env[62460]: DEBUG nova.scheduler.client.report [None req-35317be6-b47b-4771-852f-43d504ddaa7e tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 945.703683] env[62460]: DEBUG oslo_vmware.api [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313915, 'name': PowerOnVM_Task, 'duration_secs': 0.48608} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.703963] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 945.704187] env[62460]: INFO nova.compute.manager [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Took 8.47 seconds to spawn the instance on the hypervisor. [ 945.704377] env[62460]: DEBUG nova.compute.manager [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 945.705163] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51f9b58e-48ad-4679-a0ac-4bdff55eab7c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.802786] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Lock "003191d3-fe38-4bde-8d7c-75b4aad413b4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.527s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.877153] env[62460]: INFO nova.compute.manager [-] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] Took 1.41 seconds to deallocate network for instance. [ 945.980258] env[62460]: DEBUG oslo_vmware.api [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313917, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.1051} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.980488] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 945.981243] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-590228f8-bda8-4a8f-ac13-f9e5d2c0c1f3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.003432] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Reconfiguring VM instance instance-0000005d to attach disk [datastore1] 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea/0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 946.003661] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-15be1334-d106-4a34-882f-1ce18fe6d373 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.021641] env[62460]: DEBUG oslo_vmware.api [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Waiting for the task: (returnval){ [ 946.021641] env[62460]: value = "task-1313918" [ 946.021641] env[62460]: _type = "Task" [ 946.021641] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.028845] env[62460]: DEBUG oslo_vmware.api [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313918, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.071837] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Preparing fetch location {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 946.072205] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Fetch image to [datastore2] OSTACK_IMG_46c5526b-d61d-43ec-82c6-82f580cb1cb2/OSTACK_IMG_46c5526b-d61d-43ec-82c6-82f580cb1cb2.vmdk {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 946.072439] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Downloading stream optimized image 99119861-894e-4add-bc8a-5414d6c25b72 to [datastore2] OSTACK_IMG_46c5526b-d61d-43ec-82c6-82f580cb1cb2/OSTACK_IMG_46c5526b-d61d-43ec-82c6-82f580cb1cb2.vmdk on the data store datastore2 as vApp {{(pid=62460) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 946.072654] env[62460]: DEBUG nova.virt.vmwareapi.images [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Downloading image file data 99119861-894e-4add-bc8a-5414d6c25b72 to the ESX as VM named 'OSTACK_IMG_46c5526b-d61d-43ec-82c6-82f580cb1cb2' {{(pid=62460) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 946.091709] env[62460]: DEBUG nova.compute.manager [req-ff993c50-408b-49bd-84e7-09a5197ab423 req-80212bdd-e06d-4692-aa1a-0507dbf36bc3 service nova] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] Received event network-vif-deleted-5102799a-f429-4aa9-aad4-e476aacddf8b {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 946.144531] env[62460]: DEBUG oslo_concurrency.lockutils [None req-35317be6-b47b-4771-852f-43d504ddaa7e tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.750s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.147670] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0f0685de-d89c-4234-9b4d-c36872e0a4f9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 1.197s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.149685] env[62460]: DEBUG oslo_vmware.rw_handles [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 946.149685] env[62460]: value = "resgroup-9" [ 946.149685] env[62460]: _type = "ResourcePool" [ 946.149685] env[62460]: }. {{(pid=62460) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 946.150425] env[62460]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-d1b0ed5f-1001-4b4a-8342-b4ee0156c522 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.170218] env[62460]: INFO nova.scheduler.client.report [None req-35317be6-b47b-4771-852f-43d504ddaa7e tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Deleted allocations for instance 6a5934ed-30dc-4512-8cf2-1e4d9f20836c [ 946.176436] env[62460]: DEBUG oslo_vmware.rw_handles [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lease: (returnval){ [ 946.176436] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]5282f591-9013-f63b-1e4e-dcf54d99f113" [ 946.176436] env[62460]: _type = "HttpNfcLease" [ 946.176436] env[62460]: } obtained for vApp import into resource pool (val){ [ 946.176436] env[62460]: value = "resgroup-9" [ 946.176436] env[62460]: _type = "ResourcePool" [ 946.176436] env[62460]: }. {{(pid=62460) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 946.176761] env[62460]: DEBUG oslo_vmware.api [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the lease: (returnval){ [ 946.176761] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]5282f591-9013-f63b-1e4e-dcf54d99f113" [ 946.176761] env[62460]: _type = "HttpNfcLease" [ 946.176761] env[62460]: } to be ready. {{(pid=62460) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 946.184678] env[62460]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 946.184678] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]5282f591-9013-f63b-1e4e-dcf54d99f113" [ 946.184678] env[62460]: _type = "HttpNfcLease" [ 946.184678] env[62460]: } is initializing. {{(pid=62460) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 946.219960] env[62460]: INFO nova.compute.manager [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Took 21.88 seconds to build instance. [ 946.384211] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4fe60fd1-0c1e-413d-b86f-9f13b8083746 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.532109] env[62460]: DEBUG oslo_vmware.api [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313918, 'name': ReconfigVM_Task, 'duration_secs': 0.333891} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.532477] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Reconfigured VM instance instance-0000005d to attach disk [datastore1] 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea/0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 946.533194] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b44e5c3f-6af6-43c9-9f63-2b083a4fa986 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.539466] env[62460]: DEBUG oslo_vmware.api [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Waiting for the task: (returnval){ [ 946.539466] env[62460]: value = "task-1313920" [ 946.539466] env[62460]: _type = "Task" [ 946.539466] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.547878] env[62460]: DEBUG oslo_vmware.api [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313920, 'name': Rename_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.677976] env[62460]: DEBUG oslo_concurrency.lockutils [None req-35317be6-b47b-4771-852f-43d504ddaa7e tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "6a5934ed-30dc-4512-8cf2-1e4d9f20836c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.440s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.688672] env[62460]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 946.688672] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]5282f591-9013-f63b-1e4e-dcf54d99f113" [ 946.688672] env[62460]: _type = "HttpNfcLease" [ 946.688672] env[62460]: } is initializing. {{(pid=62460) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 946.721904] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ef66d39-def6-4ecf-b87d-e0e2ed1fb88d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Lock "db331001-1744-4c42-a40c-87f3e60da4b8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.414s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.794700] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17e45ca5-e119-47c0-b6a6-f6b400c151fb {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.802324] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-709b4e50-bed5-48a4-81ec-86d3ddd044bc {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.836445] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94cfb896-6a08-43c5-9e77-50a655e6542a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.844007] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d28280a2-a355-4340-b87b-ba4f712e1f0f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.857110] env[62460]: DEBUG nova.compute.provider_tree [None req-0f0685de-d89c-4234-9b4d-c36872e0a4f9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 947.052603] env[62460]: DEBUG oslo_vmware.api [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313920, 'name': Rename_Task, 'duration_secs': 0.192184} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.053570] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 947.053724] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5c77b1b3-5ae3-4b14-9ff8-63aeba8a5501 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.061867] env[62460]: DEBUG oslo_vmware.api [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Waiting for the task: (returnval){ [ 947.061867] env[62460]: value = "task-1313921" [ 947.061867] env[62460]: _type = "Task" [ 947.061867] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.070008] env[62460]: DEBUG oslo_vmware.api [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313921, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.188237] env[62460]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 947.188237] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]5282f591-9013-f63b-1e4e-dcf54d99f113" [ 947.188237] env[62460]: _type = "HttpNfcLease" [ 947.188237] env[62460]: } is ready. {{(pid=62460) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 947.188960] env[62460]: DEBUG oslo_vmware.rw_handles [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 947.188960] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]5282f591-9013-f63b-1e4e-dcf54d99f113" [ 947.188960] env[62460]: _type = "HttpNfcLease" [ 947.188960] env[62460]: }. {{(pid=62460) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 947.190051] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8616d442-41f9-43ca-abbd-9d7b92742421 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.197797] env[62460]: DEBUG oslo_vmware.rw_handles [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520cb7af-2ac3-3fa0-d730-7e19e8e2ffa8/disk-0.vmdk from lease info. {{(pid=62460) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 947.197797] env[62460]: DEBUG oslo_vmware.rw_handles [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Creating HTTP connection to write to file with size = 31669248 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520cb7af-2ac3-3fa0-d730-7e19e8e2ffa8/disk-0.vmdk. {{(pid=62460) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 947.268077] env[62460]: DEBUG oslo_concurrency.lockutils [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "290500b5-d5b4-43b6-a843-20dacf561f4b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.268355] env[62460]: DEBUG oslo_concurrency.lockutils [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "290500b5-d5b4-43b6-a843-20dacf561f4b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.273405] env[62460]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-d112daba-c26f-4c0a-a66c-0a4a4b46bbdf {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.378358] env[62460]: ERROR nova.scheduler.client.report [None req-0f0685de-d89c-4234-9b4d-c36872e0a4f9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [req-cc3089bf-f18f-41b4-a08c-b3716900e588] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 76e8f00d-65cc-4766-bb9d-5006fb51c728. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-cc3089bf-f18f-41b4-a08c-b3716900e588"}]} [ 947.397122] env[62460]: DEBUG nova.scheduler.client.report [None req-0f0685de-d89c-4234-9b4d-c36872e0a4f9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Refreshing inventories for resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 947.413133] env[62460]: DEBUG nova.scheduler.client.report [None req-0f0685de-d89c-4234-9b4d-c36872e0a4f9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Updating ProviderTree inventory for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 947.413406] env[62460]: DEBUG nova.compute.provider_tree [None req-0f0685de-d89c-4234-9b4d-c36872e0a4f9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 947.425792] env[62460]: DEBUG nova.scheduler.client.report [None req-0f0685de-d89c-4234-9b4d-c36872e0a4f9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Refreshing aggregate associations for resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728, aggregates: None {{(pid=62460) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 947.447729] env[62460]: DEBUG nova.scheduler.client.report [None req-0f0685de-d89c-4234-9b4d-c36872e0a4f9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Refreshing trait associations for resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62460) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 947.547207] env[62460]: DEBUG oslo_concurrency.lockutils [None req-f30c4799-8957-4609-a629-36257c62291c tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Acquiring lock "9c63946c-f294-42ea-979e-68db26d88858" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.547491] env[62460]: DEBUG oslo_concurrency.lockutils [None req-f30c4799-8957-4609-a629-36257c62291c tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Lock "9c63946c-f294-42ea-979e-68db26d88858" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.547711] env[62460]: DEBUG oslo_concurrency.lockutils [None req-f30c4799-8957-4609-a629-36257c62291c tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Acquiring lock "9c63946c-f294-42ea-979e-68db26d88858-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.547901] env[62460]: DEBUG oslo_concurrency.lockutils [None req-f30c4799-8957-4609-a629-36257c62291c tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Lock "9c63946c-f294-42ea-979e-68db26d88858-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.548130] env[62460]: DEBUG oslo_concurrency.lockutils [None req-f30c4799-8957-4609-a629-36257c62291c tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Lock "9c63946c-f294-42ea-979e-68db26d88858-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.551973] env[62460]: INFO nova.compute.manager [None req-f30c4799-8957-4609-a629-36257c62291c tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Terminating instance [ 947.556783] env[62460]: DEBUG nova.compute.manager [None req-f30c4799-8957-4609-a629-36257c62291c tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 947.556783] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-f30c4799-8957-4609-a629-36257c62291c tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 947.557358] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e235c70c-edf8-402b-b38f-64ae4c101306 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.567792] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-f30c4799-8957-4609-a629-36257c62291c tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 947.571966] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-91376126-5aec-4a79-a0d8-d80964093c0d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.577423] env[62460]: DEBUG oslo_vmware.api [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313921, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.583100] env[62460]: DEBUG oslo_vmware.api [None req-f30c4799-8957-4609-a629-36257c62291c tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Waiting for the task: (returnval){ [ 947.583100] env[62460]: value = "task-1313922" [ 947.583100] env[62460]: _type = "Task" [ 947.583100] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.598895] env[62460]: DEBUG oslo_vmware.api [None req-f30c4799-8957-4609-a629-36257c62291c tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313922, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.647516] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da34a280-d623-401d-a652-64ea47c7a799 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.657904] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d169fd27-c7e9-44b9-9d9d-e0f5c2468be3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.694103] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61bd0f45-c593-45fd-afd4-47e49853161c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.704573] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64a5ed08-1b1b-42de-91dc-3f7b46575654 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.721029] env[62460]: DEBUG nova.compute.provider_tree [None req-0f0685de-d89c-4234-9b4d-c36872e0a4f9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 947.773954] env[62460]: DEBUG nova.compute.manager [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 948.076985] env[62460]: DEBUG oslo_vmware.api [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313921, 'name': PowerOnVM_Task, 'duration_secs': 0.562183} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.077397] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 948.077700] env[62460]: INFO nova.compute.manager [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Took 8.53 seconds to spawn the instance on the hypervisor. [ 948.077985] env[62460]: DEBUG nova.compute.manager [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 948.079126] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66ba9dbb-394a-48f3-9276-13ef8380a141 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.100575] env[62460]: DEBUG oslo_vmware.api [None req-f30c4799-8957-4609-a629-36257c62291c tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313922, 'name': PowerOffVM_Task, 'duration_secs': 0.271002} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.100798] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-f30c4799-8957-4609-a629-36257c62291c tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 948.100977] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-f30c4799-8957-4609-a629-36257c62291c tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 948.101245] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8d6c434a-2219-436f-a6ae-1c558b93ad3c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.173707] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-f30c4799-8957-4609-a629-36257c62291c tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 948.173935] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-f30c4799-8957-4609-a629-36257c62291c tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Deleting contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 948.174253] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-f30c4799-8957-4609-a629-36257c62291c tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Deleting the datastore file [datastore1] 9c63946c-f294-42ea-979e-68db26d88858 {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 948.174546] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b53aa35c-7128-4bec-8296-17e3e21e4b51 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.183195] env[62460]: DEBUG oslo_vmware.api [None req-f30c4799-8957-4609-a629-36257c62291c tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Waiting for the task: (returnval){ [ 948.183195] env[62460]: value = "task-1313924" [ 948.183195] env[62460]: _type = "Task" [ 948.183195] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.194444] env[62460]: DEBUG oslo_vmware.api [None req-f30c4799-8957-4609-a629-36257c62291c tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313924, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.260064] env[62460]: DEBUG nova.scheduler.client.report [None req-0f0685de-d89c-4234-9b4d-c36872e0a4f9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Updated inventory for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with generation 114 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 948.260496] env[62460]: DEBUG nova.compute.provider_tree [None req-0f0685de-d89c-4234-9b4d-c36872e0a4f9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Updating resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 generation from 114 to 115 during operation: update_inventory {{(pid=62460) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 948.260858] env[62460]: DEBUG nova.compute.provider_tree [None req-0f0685de-d89c-4234-9b4d-c36872e0a4f9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 948.301206] env[62460]: DEBUG oslo_concurrency.lockutils [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.575542] env[62460]: DEBUG oslo_vmware.rw_handles [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Completed reading data from the image iterator. {{(pid=62460) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 948.575786] env[62460]: DEBUG oslo_vmware.rw_handles [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520cb7af-2ac3-3fa0-d730-7e19e8e2ffa8/disk-0.vmdk. {{(pid=62460) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 948.577001] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7490f37-4a95-4ce0-9679-935582b5cfdf {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.584235] env[62460]: DEBUG oslo_vmware.rw_handles [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520cb7af-2ac3-3fa0-d730-7e19e8e2ffa8/disk-0.vmdk is in state: ready. {{(pid=62460) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 948.584527] env[62460]: DEBUG oslo_vmware.rw_handles [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520cb7af-2ac3-3fa0-d730-7e19e8e2ffa8/disk-0.vmdk. {{(pid=62460) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 948.584825] env[62460]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-68fce811-4066-4e12-8034-ac1b3d381e21 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.605870] env[62460]: INFO nova.compute.manager [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Took 20.71 seconds to build instance. [ 948.711308] env[62460]: DEBUG oslo_vmware.api [None req-f30c4799-8957-4609-a629-36257c62291c tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313924, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.179914} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.711707] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-f30c4799-8957-4609-a629-36257c62291c tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 948.712027] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-f30c4799-8957-4609-a629-36257c62291c tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Deleted contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 948.712334] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-f30c4799-8957-4609-a629-36257c62291c tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 948.712628] env[62460]: INFO nova.compute.manager [None req-f30c4799-8957-4609-a629-36257c62291c tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Took 1.16 seconds to destroy the instance on the hypervisor. [ 948.713019] env[62460]: DEBUG oslo.service.loopingcall [None req-f30c4799-8957-4609-a629-36257c62291c tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 948.713324] env[62460]: DEBUG nova.compute.manager [-] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 948.713478] env[62460]: DEBUG nova.network.neutron [-] [instance: 9c63946c-f294-42ea-979e-68db26d88858] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 948.997331] env[62460]: DEBUG nova.compute.manager [req-1e57863a-7ffd-40c5-bf2b-9199e84cd252 req-b1eb32d2-8a83-4b29-baf4-305575b0cd9a service nova] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Received event network-vif-deleted-2770c44d-3afd-4465-84ff-de06bb9b1a18 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 948.997529] env[62460]: INFO nova.compute.manager [req-1e57863a-7ffd-40c5-bf2b-9199e84cd252 req-b1eb32d2-8a83-4b29-baf4-305575b0cd9a service nova] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Neutron deleted interface 2770c44d-3afd-4465-84ff-de06bb9b1a18; detaching it from the instance and deleting it from the info cache [ 948.998500] env[62460]: DEBUG nova.network.neutron [req-1e57863a-7ffd-40c5-bf2b-9199e84cd252 req-b1eb32d2-8a83-4b29-baf4-305575b0cd9a service nova] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 949.002206] env[62460]: DEBUG oslo_vmware.rw_handles [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520cb7af-2ac3-3fa0-d730-7e19e8e2ffa8/disk-0.vmdk. {{(pid=62460) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 949.002440] env[62460]: INFO nova.virt.vmwareapi.images [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Downloaded image file data 99119861-894e-4add-bc8a-5414d6c25b72 [ 949.003693] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdfa5a69-b2c0-4d12-8496-f45d84fbc0fd {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.023654] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-08f5d000-a2f2-46d5-87cd-58cfb6d55fb6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.108792] env[62460]: DEBUG oslo_concurrency.lockutils [None req-87a97320-a5dc-4f9a-ac1f-5f094edf945a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.227s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.212495] env[62460]: INFO nova.virt.vmwareapi.images [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] The imported VM was unregistered [ 949.214955] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Caching image {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 949.215220] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Creating directory with path [datastore2] devstack-image-cache_base/99119861-894e-4add-bc8a-5414d6c25b72 {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 949.215505] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-24dd4a21-0127-4d31-b04e-200da6cd2ac5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.228040] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Created directory with path [datastore2] devstack-image-cache_base/99119861-894e-4add-bc8a-5414d6c25b72 {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 949.228274] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_46c5526b-d61d-43ec-82c6-82f580cb1cb2/OSTACK_IMG_46c5526b-d61d-43ec-82c6-82f580cb1cb2.vmdk to [datastore2] devstack-image-cache_base/99119861-894e-4add-bc8a-5414d6c25b72/99119861-894e-4add-bc8a-5414d6c25b72.vmdk. {{(pid=62460) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 949.229141] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-b0365da5-6599-4697-b94a-b1ce538b594c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.236673] env[62460]: DEBUG oslo_vmware.api [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 949.236673] env[62460]: value = "task-1313926" [ 949.236673] env[62460]: _type = "Task" [ 949.236673] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.246503] env[62460]: DEBUG oslo_vmware.api [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313926, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.273377] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0f0685de-d89c-4234-9b4d-c36872e0a4f9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 3.127s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.276334] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4fe60fd1-0c1e-413d-b86f-9f13b8083746 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.892s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.276601] env[62460]: DEBUG nova.objects.instance [None req-4fe60fd1-0c1e-413d-b86f-9f13b8083746 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lazy-loading 'resources' on Instance uuid e90e8c3a-d244-4d04-997e-c15ab1cb2009 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 949.461240] env[62460]: DEBUG nova.network.neutron [-] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 949.500610] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-64473fac-cfa5-468e-9202-14dc59d5a8cf {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.509750] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcfbe021-e398-4285-a606-94f98b69a34d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.536468] env[62460]: DEBUG nova.compute.manager [req-1e57863a-7ffd-40c5-bf2b-9199e84cd252 req-b1eb32d2-8a83-4b29-baf4-305575b0cd9a service nova] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Detach interface failed, port_id=2770c44d-3afd-4465-84ff-de06bb9b1a18, reason: Instance 9c63946c-f294-42ea-979e-68db26d88858 could not be found. {{(pid=62460) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 949.746950] env[62460]: DEBUG oslo_vmware.api [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313926, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.792407] env[62460]: DEBUG nova.compute.manager [req-fa882f17-f92d-4fd2-92ed-b38c45c9b429 req-86636fd0-c571-4ff1-8d79-0a774f6159b3 service nova] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Received event network-changed-546b0347-51db-4a4a-99e9-4f3db96a36bd {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 949.792617] env[62460]: DEBUG nova.compute.manager [req-fa882f17-f92d-4fd2-92ed-b38c45c9b429 req-86636fd0-c571-4ff1-8d79-0a774f6159b3 service nova] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Refreshing instance network info cache due to event network-changed-546b0347-51db-4a4a-99e9-4f3db96a36bd. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 949.792861] env[62460]: DEBUG oslo_concurrency.lockutils [req-fa882f17-f92d-4fd2-92ed-b38c45c9b429 req-86636fd0-c571-4ff1-8d79-0a774f6159b3 service nova] Acquiring lock "refresh_cache-abd4501b-3ad8-4103-921d-5b80fc2f313f" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 949.793088] env[62460]: DEBUG oslo_concurrency.lockutils [req-fa882f17-f92d-4fd2-92ed-b38c45c9b429 req-86636fd0-c571-4ff1-8d79-0a774f6159b3 service nova] Acquired lock "refresh_cache-abd4501b-3ad8-4103-921d-5b80fc2f313f" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.793271] env[62460]: DEBUG nova.network.neutron [req-fa882f17-f92d-4fd2-92ed-b38c45c9b429 req-86636fd0-c571-4ff1-8d79-0a774f6159b3 service nova] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Refreshing network info cache for port 546b0347-51db-4a4a-99e9-4f3db96a36bd {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 949.855265] env[62460]: INFO nova.scheduler.client.report [None req-0f0685de-d89c-4234-9b4d-c36872e0a4f9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Deleted allocation for migration 68cc2b47-75ac-432c-be4d-b2f8553dfe61 [ 949.960068] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a926a5ed-4bd0-4698-a2f5-4601a127a932 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.963567] env[62460]: INFO nova.compute.manager [-] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Took 1.25 seconds to deallocate network for instance. [ 949.975346] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbd47d7c-a5b3-4125-9c0c-1d28f2ed687c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.015141] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-968d20bd-8856-43bd-9d2a-30d0e059d0fe {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.027277] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bed4c9c3-bb18-4971-a2d2-9c70cd545bb7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.044333] env[62460]: DEBUG nova.compute.provider_tree [None req-4fe60fd1-0c1e-413d-b86f-9f13b8083746 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 950.251496] env[62460]: DEBUG oslo_vmware.api [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313926, 'name': MoveVirtualDisk_Task} progress is 15%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.359800] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0f0685de-d89c-4234-9b4d-c36872e0a4f9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "6b084ed4-e8f0-4063-bf2a-252740640753" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 8.948s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.483503] env[62460]: DEBUG oslo_concurrency.lockutils [None req-f30c4799-8957-4609-a629-36257c62291c tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.548484] env[62460]: DEBUG nova.scheduler.client.report [None req-4fe60fd1-0c1e-413d-b86f-9f13b8083746 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 950.618053] env[62460]: DEBUG nova.network.neutron [req-fa882f17-f92d-4fd2-92ed-b38c45c9b429 req-86636fd0-c571-4ff1-8d79-0a774f6159b3 service nova] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Updated VIF entry in instance network info cache for port 546b0347-51db-4a4a-99e9-4f3db96a36bd. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 950.618226] env[62460]: DEBUG nova.network.neutron [req-fa882f17-f92d-4fd2-92ed-b38c45c9b429 req-86636fd0-c571-4ff1-8d79-0a774f6159b3 service nova] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Updating instance_info_cache with network_info: [{"id": "546b0347-51db-4a4a-99e9-4f3db96a36bd", "address": "fa:16:3e:54:9b:7b", "network": {"id": "d3136e32-ad55-4b73-835c-8fa5f0480767", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1511414968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53fa06008e9f43488362895e7a143700", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "054fcd1e-638e-425a-a1de-78cb188ae026", "external-id": "nsx-vlan-transportzone-658", "segmentation_id": 658, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap546b0347-51", "ovs_interfaceid": "546b0347-51db-4a4a-99e9-4f3db96a36bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.751892] env[62460]: DEBUG oslo_vmware.api [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313926, 'name': MoveVirtualDisk_Task} progress is 38%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.057613] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4fe60fd1-0c1e-413d-b86f-9f13b8083746 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.781s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.061348] env[62460]: DEBUG oslo_concurrency.lockutils [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.760s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.063407] env[62460]: INFO nova.compute.claims [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 951.090510] env[62460]: INFO nova.scheduler.client.report [None req-4fe60fd1-0c1e-413d-b86f-9f13b8083746 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Deleted allocations for instance e90e8c3a-d244-4d04-997e-c15ab1cb2009 [ 951.121795] env[62460]: DEBUG oslo_concurrency.lockutils [req-fa882f17-f92d-4fd2-92ed-b38c45c9b429 req-86636fd0-c571-4ff1-8d79-0a774f6159b3 service nova] Releasing lock "refresh_cache-abd4501b-3ad8-4103-921d-5b80fc2f313f" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.252878] env[62460]: DEBUG oslo_vmware.api [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313926, 'name': MoveVirtualDisk_Task} progress is 57%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.598433] env[62460]: DEBUG oslo_concurrency.lockutils [None req-4fe60fd1-0c1e-413d-b86f-9f13b8083746 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "e90e8c3a-d244-4d04-997e-c15ab1cb2009" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.296s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.643874] env[62460]: DEBUG nova.compute.manager [req-d73f42cb-60bd-413d-b515-8f0f57c11d47 req-8b647065-4f22-4d7b-a59c-3e1293c7faf6 service nova] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Received event network-changed-90a7d774-4d76-4525-8e6d-8c08fdb93fb4 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 951.643874] env[62460]: DEBUG nova.compute.manager [req-d73f42cb-60bd-413d-b515-8f0f57c11d47 req-8b647065-4f22-4d7b-a59c-3e1293c7faf6 service nova] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Refreshing instance network info cache due to event network-changed-90a7d774-4d76-4525-8e6d-8c08fdb93fb4. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 951.643874] env[62460]: DEBUG oslo_concurrency.lockutils [req-d73f42cb-60bd-413d-b515-8f0f57c11d47 req-8b647065-4f22-4d7b-a59c-3e1293c7faf6 service nova] Acquiring lock "refresh_cache-0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.643874] env[62460]: DEBUG oslo_concurrency.lockutils [req-d73f42cb-60bd-413d-b515-8f0f57c11d47 req-8b647065-4f22-4d7b-a59c-3e1293c7faf6 service nova] Acquired lock "refresh_cache-0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.643874] env[62460]: DEBUG nova.network.neutron [req-d73f42cb-60bd-413d-b515-8f0f57c11d47 req-8b647065-4f22-4d7b-a59c-3e1293c7faf6 service nova] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Refreshing network info cache for port 90a7d774-4d76-4525-8e6d-8c08fdb93fb4 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 951.751802] env[62460]: DEBUG oslo_vmware.api [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313926, 'name': MoveVirtualDisk_Task} progress is 80%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.892778] env[62460]: DEBUG nova.compute.manager [req-e79d00cf-e61e-4791-bb97-8186903fe1be req-5adecb9f-a8f5-4641-8038-5c0e85879ad8 service nova] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Received event network-changed-90a7d774-4d76-4525-8e6d-8c08fdb93fb4 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 951.892778] env[62460]: DEBUG nova.compute.manager [req-e79d00cf-e61e-4791-bb97-8186903fe1be req-5adecb9f-a8f5-4641-8038-5c0e85879ad8 service nova] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Refreshing instance network info cache due to event network-changed-90a7d774-4d76-4525-8e6d-8c08fdb93fb4. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 951.892993] env[62460]: DEBUG oslo_concurrency.lockutils [req-e79d00cf-e61e-4791-bb97-8186903fe1be req-5adecb9f-a8f5-4641-8038-5c0e85879ad8 service nova] Acquiring lock "refresh_cache-0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.233232] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccaa56c9-6933-4d1b-a881-0f4caf8ca8c6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.249767] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3c0af58-d7fa-4bd1-9c3b-c2e56d7c9f93 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.259584] env[62460]: DEBUG oslo_vmware.api [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313926, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.290525] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-713d48fc-dcb3-4d0b-bbc6-3f1fbf91a508 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.298456] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34d3c58c-9292-4a32-bdaf-47590905274c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.313178] env[62460]: DEBUG nova.compute.provider_tree [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 952.473890] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d1afac0f-2ef1-496a-b856-f36557e80c73 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "6b084ed4-e8f0-4063-bf2a-252740640753" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.473890] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d1afac0f-2ef1-496a-b856-f36557e80c73 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "6b084ed4-e8f0-4063-bf2a-252740640753" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.474436] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d1afac0f-2ef1-496a-b856-f36557e80c73 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "6b084ed4-e8f0-4063-bf2a-252740640753-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.474436] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d1afac0f-2ef1-496a-b856-f36557e80c73 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "6b084ed4-e8f0-4063-bf2a-252740640753-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.474558] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d1afac0f-2ef1-496a-b856-f36557e80c73 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "6b084ed4-e8f0-4063-bf2a-252740640753-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.476940] env[62460]: DEBUG nova.network.neutron [req-d73f42cb-60bd-413d-b515-8f0f57c11d47 req-8b647065-4f22-4d7b-a59c-3e1293c7faf6 service nova] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Updated VIF entry in instance network info cache for port 90a7d774-4d76-4525-8e6d-8c08fdb93fb4. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 952.477302] env[62460]: DEBUG nova.network.neutron [req-d73f42cb-60bd-413d-b515-8f0f57c11d47 req-8b647065-4f22-4d7b-a59c-3e1293c7faf6 service nova] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Updating instance_info_cache with network_info: [{"id": "90a7d774-4d76-4525-8e6d-8c08fdb93fb4", "address": "fa:16:3e:5b:1c:b7", "network": {"id": "d3136e32-ad55-4b73-835c-8fa5f0480767", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1511414968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53fa06008e9f43488362895e7a143700", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "054fcd1e-638e-425a-a1de-78cb188ae026", "external-id": "nsx-vlan-transportzone-658", "segmentation_id": 658, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90a7d774-4d", "ovs_interfaceid": "90a7d774-4d76-4525-8e6d-8c08fdb93fb4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.479055] env[62460]: INFO nova.compute.manager [None req-d1afac0f-2ef1-496a-b856-f36557e80c73 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Terminating instance [ 952.480768] env[62460]: DEBUG nova.compute.manager [None req-d1afac0f-2ef1-496a-b856-f36557e80c73 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 952.480912] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-d1afac0f-2ef1-496a-b856-f36557e80c73 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 952.481765] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28be6293-cdef-4ea5-86a4-cfe70b33e798 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.489200] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1afac0f-2ef1-496a-b856-f36557e80c73 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 952.489991] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-50614847-fdde-46d5-8237-b1cf4d8af917 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.496632] env[62460]: DEBUG oslo_vmware.api [None req-d1afac0f-2ef1-496a-b856-f36557e80c73 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 952.496632] env[62460]: value = "task-1313927" [ 952.496632] env[62460]: _type = "Task" [ 952.496632] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.504843] env[62460]: DEBUG oslo_vmware.api [None req-d1afac0f-2ef1-496a-b856-f36557e80c73 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313927, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.651728] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "f0f66557-04ef-4a96-8a56-e640bfda3b36" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.652076] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "f0f66557-04ef-4a96-8a56-e640bfda3b36" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.749728] env[62460]: DEBUG oslo_vmware.api [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313926, 'name': MoveVirtualDisk_Task, 'duration_secs': 3.207105} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.749962] env[62460]: INFO nova.virt.vmwareapi.ds_util [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_46c5526b-d61d-43ec-82c6-82f580cb1cb2/OSTACK_IMG_46c5526b-d61d-43ec-82c6-82f580cb1cb2.vmdk to [datastore2] devstack-image-cache_base/99119861-894e-4add-bc8a-5414d6c25b72/99119861-894e-4add-bc8a-5414d6c25b72.vmdk. [ 952.750207] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Cleaning up location [datastore2] OSTACK_IMG_46c5526b-d61d-43ec-82c6-82f580cb1cb2 {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 952.750388] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_46c5526b-d61d-43ec-82c6-82f580cb1cb2 {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 952.750642] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5317ea5c-6f2d-4741-bdf8-4e496f1c8604 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.756498] env[62460]: DEBUG oslo_vmware.api [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 952.756498] env[62460]: value = "task-1313928" [ 952.756498] env[62460]: _type = "Task" [ 952.756498] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.764132] env[62460]: DEBUG oslo_vmware.api [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313928, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.816513] env[62460]: DEBUG nova.scheduler.client.report [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 952.981935] env[62460]: DEBUG oslo_concurrency.lockutils [req-d73f42cb-60bd-413d-b515-8f0f57c11d47 req-8b647065-4f22-4d7b-a59c-3e1293c7faf6 service nova] Releasing lock "refresh_cache-0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.982414] env[62460]: DEBUG oslo_concurrency.lockutils [req-e79d00cf-e61e-4791-bb97-8186903fe1be req-5adecb9f-a8f5-4641-8038-5c0e85879ad8 service nova] Acquired lock "refresh_cache-0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.982664] env[62460]: DEBUG nova.network.neutron [req-e79d00cf-e61e-4791-bb97-8186903fe1be req-5adecb9f-a8f5-4641-8038-5c0e85879ad8 service nova] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Refreshing network info cache for port 90a7d774-4d76-4525-8e6d-8c08fdb93fb4 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 953.006282] env[62460]: DEBUG oslo_vmware.api [None req-d1afac0f-2ef1-496a-b856-f36557e80c73 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313927, 'name': PowerOffVM_Task, 'duration_secs': 0.488073} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.006456] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1afac0f-2ef1-496a-b856-f36557e80c73 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 953.006633] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-d1afac0f-2ef1-496a-b856-f36557e80c73 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 953.006887] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4e18fd3c-8385-494c-a8dc-0b53e4bac61b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.041646] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "0d39a304-6a49-4d87-bfa2-4ba3b383578a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.041780] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "0d39a304-6a49-4d87-bfa2-4ba3b383578a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.094036] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-d1afac0f-2ef1-496a-b856-f36557e80c73 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 953.095108] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-d1afac0f-2ef1-496a-b856-f36557e80c73 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Deleting contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 953.095108] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1afac0f-2ef1-496a-b856-f36557e80c73 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Deleting the datastore file [datastore2] 6b084ed4-e8f0-4063-bf2a-252740640753 {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 953.095108] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f6d07593-c748-4d55-b5cc-f39e9780dd53 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.103047] env[62460]: DEBUG oslo_vmware.api [None req-d1afac0f-2ef1-496a-b856-f36557e80c73 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 953.103047] env[62460]: value = "task-1313930" [ 953.103047] env[62460]: _type = "Task" [ 953.103047] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.110633] env[62460]: DEBUG oslo_vmware.api [None req-d1afac0f-2ef1-496a-b856-f36557e80c73 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313930, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.156061] env[62460]: DEBUG nova.compute.manager [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 953.267443] env[62460]: DEBUG oslo_vmware.api [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313928, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.047389} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.267743] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 953.267892] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Releasing lock "[datastore2] devstack-image-cache_base/99119861-894e-4add-bc8a-5414d6c25b72/99119861-894e-4add-bc8a-5414d6c25b72.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.268189] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/99119861-894e-4add-bc8a-5414d6c25b72/99119861-894e-4add-bc8a-5414d6c25b72.vmdk to [datastore2] 896f8ff4-e45f-4403-a727-03ee25e58609/896f8ff4-e45f-4403-a727-03ee25e58609.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 953.268523] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0150e7e5-daae-4d65-96c6-57a1458b06bc {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.275745] env[62460]: DEBUG oslo_vmware.api [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 953.275745] env[62460]: value = "task-1313931" [ 953.275745] env[62460]: _type = "Task" [ 953.275745] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.284136] env[62460]: DEBUG oslo_vmware.api [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313931, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.320902] env[62460]: DEBUG oslo_concurrency.lockutils [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.260s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.321487] env[62460]: DEBUG nova.compute.manager [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 953.324072] env[62460]: DEBUG oslo_concurrency.lockutils [None req-f30c4799-8957-4609-a629-36257c62291c tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.841s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.324309] env[62460]: DEBUG nova.objects.instance [None req-f30c4799-8957-4609-a629-36257c62291c tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Lazy-loading 'resources' on Instance uuid 9c63946c-f294-42ea-979e-68db26d88858 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 953.544978] env[62460]: DEBUG nova.compute.manager [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 953.612688] env[62460]: DEBUG oslo_vmware.api [None req-d1afac0f-2ef1-496a-b856-f36557e80c73 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313930, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146938} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.612987] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1afac0f-2ef1-496a-b856-f36557e80c73 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 953.613202] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-d1afac0f-2ef1-496a-b856-f36557e80c73 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Deleted contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 953.613389] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-d1afac0f-2ef1-496a-b856-f36557e80c73 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 953.613572] env[62460]: INFO nova.compute.manager [None req-d1afac0f-2ef1-496a-b856-f36557e80c73 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Took 1.13 seconds to destroy the instance on the hypervisor. [ 953.613827] env[62460]: DEBUG oslo.service.loopingcall [None req-d1afac0f-2ef1-496a-b856-f36557e80c73 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 953.614048] env[62460]: DEBUG nova.compute.manager [-] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 953.614151] env[62460]: DEBUG nova.network.neutron [-] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 953.678559] env[62460]: DEBUG nova.compute.manager [req-8f992b59-2445-426b-909d-f2e3d425d291 req-a932dd3a-4a47-47a1-9385-693e5c901c6b service nova] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Received event network-changed-546b0347-51db-4a4a-99e9-4f3db96a36bd {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 953.679012] env[62460]: DEBUG nova.compute.manager [req-8f992b59-2445-426b-909d-f2e3d425d291 req-a932dd3a-4a47-47a1-9385-693e5c901c6b service nova] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Refreshing instance network info cache due to event network-changed-546b0347-51db-4a4a-99e9-4f3db96a36bd. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 953.679468] env[62460]: DEBUG oslo_concurrency.lockutils [req-8f992b59-2445-426b-909d-f2e3d425d291 req-a932dd3a-4a47-47a1-9385-693e5c901c6b service nova] Acquiring lock "refresh_cache-abd4501b-3ad8-4103-921d-5b80fc2f313f" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.679688] env[62460]: DEBUG oslo_concurrency.lockutils [req-8f992b59-2445-426b-909d-f2e3d425d291 req-a932dd3a-4a47-47a1-9385-693e5c901c6b service nova] Acquired lock "refresh_cache-abd4501b-3ad8-4103-921d-5b80fc2f313f" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.679890] env[62460]: DEBUG nova.network.neutron [req-8f992b59-2445-426b-909d-f2e3d425d291 req-a932dd3a-4a47-47a1-9385-693e5c901c6b service nova] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Refreshing network info cache for port 546b0347-51db-4a4a-99e9-4f3db96a36bd {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 953.684657] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.788621] env[62460]: DEBUG oslo_vmware.api [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313931, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.827595] env[62460]: DEBUG nova.compute.utils [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 953.831902] env[62460]: DEBUG nova.compute.manager [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 953.832135] env[62460]: DEBUG nova.network.neutron [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 953.842882] env[62460]: DEBUG nova.network.neutron [req-e79d00cf-e61e-4791-bb97-8186903fe1be req-5adecb9f-a8f5-4641-8038-5c0e85879ad8 service nova] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Updated VIF entry in instance network info cache for port 90a7d774-4d76-4525-8e6d-8c08fdb93fb4. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 953.842882] env[62460]: DEBUG nova.network.neutron [req-e79d00cf-e61e-4791-bb97-8186903fe1be req-5adecb9f-a8f5-4641-8038-5c0e85879ad8 service nova] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Updating instance_info_cache with network_info: [{"id": "90a7d774-4d76-4525-8e6d-8c08fdb93fb4", "address": "fa:16:3e:5b:1c:b7", "network": {"id": "d3136e32-ad55-4b73-835c-8fa5f0480767", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1511414968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53fa06008e9f43488362895e7a143700", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "054fcd1e-638e-425a-a1de-78cb188ae026", "external-id": "nsx-vlan-transportzone-658", "segmentation_id": 658, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90a7d774-4d", "ovs_interfaceid": "90a7d774-4d76-4525-8e6d-8c08fdb93fb4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.877207] env[62460]: DEBUG nova.policy [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5f69b3e3f50248ff8e57054294a04686', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0da056d93bdf40c39d6e82e457727ff6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 954.015403] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e4d51d8-be28-4c73-b00c-4bd85eee54d1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.025352] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3d7372c-2290-4cd9-99c1-5d5a516849cb {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.064145] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d410ca94-88dc-487e-9d5d-31e89dd532a1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.078353] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e77ff21e-d44a-4734-afa6-6c237b2b139a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.082685] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.094584] env[62460]: DEBUG nova.compute.provider_tree [None req-f30c4799-8957-4609-a629-36257c62291c tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 954.189367] env[62460]: DEBUG nova.network.neutron [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Successfully created port: 9f9c5b7b-922a-4387-b997-3f552dd85a50 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 954.291707] env[62460]: DEBUG oslo_vmware.api [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313931, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.332991] env[62460]: DEBUG nova.compute.manager [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 954.344722] env[62460]: DEBUG oslo_concurrency.lockutils [req-e79d00cf-e61e-4791-bb97-8186903fe1be req-5adecb9f-a8f5-4641-8038-5c0e85879ad8 service nova] Releasing lock "refresh_cache-0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.460723] env[62460]: DEBUG nova.network.neutron [-] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.534586] env[62460]: DEBUG nova.network.neutron [req-8f992b59-2445-426b-909d-f2e3d425d291 req-a932dd3a-4a47-47a1-9385-693e5c901c6b service nova] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Updated VIF entry in instance network info cache for port 546b0347-51db-4a4a-99e9-4f3db96a36bd. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 954.534977] env[62460]: DEBUG nova.network.neutron [req-8f992b59-2445-426b-909d-f2e3d425d291 req-a932dd3a-4a47-47a1-9385-693e5c901c6b service nova] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Updating instance_info_cache with network_info: [{"id": "546b0347-51db-4a4a-99e9-4f3db96a36bd", "address": "fa:16:3e:54:9b:7b", "network": {"id": "d3136e32-ad55-4b73-835c-8fa5f0480767", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1511414968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53fa06008e9f43488362895e7a143700", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "054fcd1e-638e-425a-a1de-78cb188ae026", "external-id": "nsx-vlan-transportzone-658", "segmentation_id": 658, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap546b0347-51", "ovs_interfaceid": "546b0347-51db-4a4a-99e9-4f3db96a36bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.619068] env[62460]: ERROR nova.scheduler.client.report [None req-f30c4799-8957-4609-a629-36257c62291c tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [req-0e8b90b8-0b85-44c6-a00c-cb4e691a6147] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 76e8f00d-65cc-4766-bb9d-5006fb51c728. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-0e8b90b8-0b85-44c6-a00c-cb4e691a6147"}]} [ 954.638829] env[62460]: DEBUG nova.scheduler.client.report [None req-f30c4799-8957-4609-a629-36257c62291c tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Refreshing inventories for resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 954.657147] env[62460]: DEBUG nova.scheduler.client.report [None req-f30c4799-8957-4609-a629-36257c62291c tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Updating ProviderTree inventory for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 954.657390] env[62460]: DEBUG nova.compute.provider_tree [None req-f30c4799-8957-4609-a629-36257c62291c tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 954.671302] env[62460]: DEBUG nova.scheduler.client.report [None req-f30c4799-8957-4609-a629-36257c62291c tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Refreshing aggregate associations for resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728, aggregates: None {{(pid=62460) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 954.692052] env[62460]: DEBUG nova.scheduler.client.report [None req-f30c4799-8957-4609-a629-36257c62291c tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Refreshing trait associations for resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62460) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 954.790100] env[62460]: DEBUG oslo_vmware.api [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313931, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.879184] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98dffde3-5ab1-41e6-b187-c1f66ed355d3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.891927] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abddd411-0d55-4cc1-b5e4-ef566387b937 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.926484] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-929831c5-7456-4970-9cd7-3fb89ded0df3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.936444] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0727838-8ffe-4f87-beef-ee2c9a30346f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.954059] env[62460]: DEBUG nova.compute.provider_tree [None req-f30c4799-8957-4609-a629-36257c62291c tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 954.962904] env[62460]: INFO nova.compute.manager [-] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Took 1.35 seconds to deallocate network for instance. [ 955.038347] env[62460]: DEBUG oslo_concurrency.lockutils [req-8f992b59-2445-426b-909d-f2e3d425d291 req-a932dd3a-4a47-47a1-9385-693e5c901c6b service nova] Releasing lock "refresh_cache-abd4501b-3ad8-4103-921d-5b80fc2f313f" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.293874] env[62460]: DEBUG oslo_vmware.api [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313931, 'name': CopyVirtualDisk_Task} progress is 83%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.345428] env[62460]: DEBUG nova.compute.manager [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 955.372043] env[62460]: DEBUG nova.virt.hardware [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 955.372328] env[62460]: DEBUG nova.virt.hardware [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 955.372495] env[62460]: DEBUG nova.virt.hardware [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 955.372688] env[62460]: DEBUG nova.virt.hardware [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 955.372843] env[62460]: DEBUG nova.virt.hardware [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 955.373133] env[62460]: DEBUG nova.virt.hardware [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 955.373375] env[62460]: DEBUG nova.virt.hardware [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 955.373549] env[62460]: DEBUG nova.virt.hardware [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 955.373727] env[62460]: DEBUG nova.virt.hardware [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 955.373903] env[62460]: DEBUG nova.virt.hardware [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 955.374106] env[62460]: DEBUG nova.virt.hardware [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 955.375010] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22b37a3c-d539-4609-88d8-f1fb8c3f6662 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.386761] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2daa3e1c-5c9b-499c-9379-169ff25d738b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.469788] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d1afac0f-2ef1-496a-b856-f36557e80c73 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.488242] env[62460]: DEBUG nova.scheduler.client.report [None req-f30c4799-8957-4609-a629-36257c62291c tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Updated inventory for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with generation 118 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 955.488598] env[62460]: DEBUG nova.compute.provider_tree [None req-f30c4799-8957-4609-a629-36257c62291c tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Updating resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 generation from 118 to 119 during operation: update_inventory {{(pid=62460) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 955.488851] env[62460]: DEBUG nova.compute.provider_tree [None req-f30c4799-8957-4609-a629-36257c62291c tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 955.709129] env[62460]: DEBUG nova.compute.manager [req-dfb548a5-408c-431f-a2d9-69c6f545327a req-1d7230cd-e7b1-4f8d-b0ef-778701bcc6f3 service nova] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Received event network-vif-deleted-3eaf04a4-feda-4a00-908e-d121fc775ecb {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 955.727686] env[62460]: DEBUG nova.compute.manager [req-77c2f248-5259-4376-a1d7-8d5a60b982ba req-58417831-6a36-4dbe-a2e4-0905f6fedfb5 service nova] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Received event network-vif-plugged-9f9c5b7b-922a-4387-b997-3f552dd85a50 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 955.727955] env[62460]: DEBUG oslo_concurrency.lockutils [req-77c2f248-5259-4376-a1d7-8d5a60b982ba req-58417831-6a36-4dbe-a2e4-0905f6fedfb5 service nova] Acquiring lock "290500b5-d5b4-43b6-a843-20dacf561f4b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.728196] env[62460]: DEBUG oslo_concurrency.lockutils [req-77c2f248-5259-4376-a1d7-8d5a60b982ba req-58417831-6a36-4dbe-a2e4-0905f6fedfb5 service nova] Lock "290500b5-d5b4-43b6-a843-20dacf561f4b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.728373] env[62460]: DEBUG oslo_concurrency.lockutils [req-77c2f248-5259-4376-a1d7-8d5a60b982ba req-58417831-6a36-4dbe-a2e4-0905f6fedfb5 service nova] Lock "290500b5-d5b4-43b6-a843-20dacf561f4b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.728549] env[62460]: DEBUG nova.compute.manager [req-77c2f248-5259-4376-a1d7-8d5a60b982ba req-58417831-6a36-4dbe-a2e4-0905f6fedfb5 service nova] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] No waiting events found dispatching network-vif-plugged-9f9c5b7b-922a-4387-b997-3f552dd85a50 {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 955.728718] env[62460]: WARNING nova.compute.manager [req-77c2f248-5259-4376-a1d7-8d5a60b982ba req-58417831-6a36-4dbe-a2e4-0905f6fedfb5 service nova] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Received unexpected event network-vif-plugged-9f9c5b7b-922a-4387-b997-3f552dd85a50 for instance with vm_state building and task_state spawning. [ 955.789074] env[62460]: DEBUG oslo_vmware.api [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313931, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.368031} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.789361] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/99119861-894e-4add-bc8a-5414d6c25b72/99119861-894e-4add-bc8a-5414d6c25b72.vmdk to [datastore2] 896f8ff4-e45f-4403-a727-03ee25e58609/896f8ff4-e45f-4403-a727-03ee25e58609.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 955.790157] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2a1af51-e59c-400c-a484-0322d781c642 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.811670] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Reconfiguring VM instance instance-0000004a to attach disk [datastore2] 896f8ff4-e45f-4403-a727-03ee25e58609/896f8ff4-e45f-4403-a727-03ee25e58609.vmdk or device None with type streamOptimized {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 955.812607] env[62460]: DEBUG nova.network.neutron [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Successfully updated port: 9f9c5b7b-922a-4387-b997-3f552dd85a50 {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 955.813734] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1e77d080-1832-4410-b164-d9d0c836af22 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.830147] env[62460]: DEBUG oslo_concurrency.lockutils [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "refresh_cache-290500b5-d5b4-43b6-a843-20dacf561f4b" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.830328] env[62460]: DEBUG oslo_concurrency.lockutils [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquired lock "refresh_cache-290500b5-d5b4-43b6-a843-20dacf561f4b" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.830500] env[62460]: DEBUG nova.network.neutron [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 955.838288] env[62460]: DEBUG oslo_vmware.api [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 955.838288] env[62460]: value = "task-1313932" [ 955.838288] env[62460]: _type = "Task" [ 955.838288] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.847144] env[62460]: DEBUG oslo_vmware.api [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313932, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.994539] env[62460]: DEBUG oslo_concurrency.lockutils [None req-f30c4799-8957-4609-a629-36257c62291c tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.670s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.997222] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.313s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.998847] env[62460]: INFO nova.compute.claims [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 956.015161] env[62460]: INFO nova.scheduler.client.report [None req-f30c4799-8957-4609-a629-36257c62291c tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Deleted allocations for instance 9c63946c-f294-42ea-979e-68db26d88858 [ 956.348908] env[62460]: DEBUG oslo_vmware.api [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313932, 'name': ReconfigVM_Task, 'duration_secs': 0.357452} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.349229] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Reconfigured VM instance instance-0000004a to attach disk [datastore2] 896f8ff4-e45f-4403-a727-03ee25e58609/896f8ff4-e45f-4403-a727-03ee25e58609.vmdk or device None with type streamOptimized {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 956.350669] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'device_type': 'disk', 'encrypted': False, 'encryption_options': None, 'guest_format': None, 'disk_bus': None, 'boot_index': 0, 'encryption_secret_uuid': None, 'encryption_format': None, 'size': 0, 'device_name': '/dev/sda', 'image_id': 'eb3fb456-2dfd-4d95-978f-f17699d7735d'}], 'ephemerals': [], 'block_device_mapping': [{'mount_device': '/dev/sdb', 'device_type': None, 'attachment_id': '0c9eb5ce-ef9b-4a0b-9d84-041b7913b082', 'guest_format': None, 'disk_bus': None, 'boot_index': None, 'delete_on_termination': False, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281244', 'volume_id': 'c472b891-df79-428d-a905-5349e414e7f5', 'name': 'volume-c472b891-df79-428d-a905-5349e414e7f5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '896f8ff4-e45f-4403-a727-03ee25e58609', 'attached_at': '', 'detached_at': '', 'volume_id': 'c472b891-df79-428d-a905-5349e414e7f5', 'serial': 'c472b891-df79-428d-a905-5349e414e7f5'}, 'volume_type': None}], 'swap': None} {{(pid=62460) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 956.350910] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Volume attach. Driver type: vmdk {{(pid=62460) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 956.351136] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281244', 'volume_id': 'c472b891-df79-428d-a905-5349e414e7f5', 'name': 'volume-c472b891-df79-428d-a905-5349e414e7f5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '896f8ff4-e45f-4403-a727-03ee25e58609', 'attached_at': '', 'detached_at': '', 'volume_id': 'c472b891-df79-428d-a905-5349e414e7f5', 'serial': 'c472b891-df79-428d-a905-5349e414e7f5'} {{(pid=62460) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 956.351929] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4597319-0837-43f6-b5b8-dabcfcefafc8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.368243] env[62460]: DEBUG nova.network.neutron [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 956.370460] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-295570af-5499-4005-8122-5551804d4ee9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.395333] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Reconfiguring VM instance instance-0000004a to attach disk [datastore2] volume-c472b891-df79-428d-a905-5349e414e7f5/volume-c472b891-df79-428d-a905-5349e414e7f5.vmdk or device None with type thin {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 956.397431] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f3b53396-8a30-441c-aec3-8a18916c380e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.415153] env[62460]: DEBUG oslo_vmware.api [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 956.415153] env[62460]: value = "task-1313933" [ 956.415153] env[62460]: _type = "Task" [ 956.415153] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.422964] env[62460]: DEBUG oslo_vmware.api [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313933, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.523401] env[62460]: DEBUG oslo_concurrency.lockutils [None req-f30c4799-8957-4609-a629-36257c62291c tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Lock "9c63946c-f294-42ea-979e-68db26d88858" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.976s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.536243] env[62460]: DEBUG nova.network.neutron [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Updating instance_info_cache with network_info: [{"id": "9f9c5b7b-922a-4387-b997-3f552dd85a50", "address": "fa:16:3e:78:80:d5", "network": {"id": "de789887-9cf9-4614-a7e1-dab430aaabaf", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-656400207-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0da056d93bdf40c39d6e82e457727ff6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5116f690-f825-4fee-8a47-42b073e716c5", "external-id": "nsx-vlan-transportzone-692", "segmentation_id": 692, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f9c5b7b-92", "ovs_interfaceid": "9f9c5b7b-922a-4387-b997-3f552dd85a50", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.926008] env[62460]: DEBUG oslo_vmware.api [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313933, 'name': ReconfigVM_Task, 'duration_secs': 0.330452} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.926332] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Reconfigured VM instance instance-0000004a to attach disk [datastore2] volume-c472b891-df79-428d-a905-5349e414e7f5/volume-c472b891-df79-428d-a905-5349e414e7f5.vmdk or device None with type thin {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 956.931109] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5489d577-d7fb-4cd5-86a0-e60235aa2b9a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.946741] env[62460]: DEBUG oslo_vmware.api [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 956.946741] env[62460]: value = "task-1313934" [ 956.946741] env[62460]: _type = "Task" [ 956.946741] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.955018] env[62460]: DEBUG oslo_vmware.api [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313934, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.038795] env[62460]: DEBUG oslo_concurrency.lockutils [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Releasing lock "refresh_cache-290500b5-d5b4-43b6-a843-20dacf561f4b" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.039234] env[62460]: DEBUG nova.compute.manager [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Instance network_info: |[{"id": "9f9c5b7b-922a-4387-b997-3f552dd85a50", "address": "fa:16:3e:78:80:d5", "network": {"id": "de789887-9cf9-4614-a7e1-dab430aaabaf", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-656400207-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0da056d93bdf40c39d6e82e457727ff6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5116f690-f825-4fee-8a47-42b073e716c5", "external-id": "nsx-vlan-transportzone-692", "segmentation_id": 692, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f9c5b7b-92", "ovs_interfaceid": "9f9c5b7b-922a-4387-b997-3f552dd85a50", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 957.039673] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:78:80:d5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5116f690-f825-4fee-8a47-42b073e716c5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9f9c5b7b-922a-4387-b997-3f552dd85a50', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 957.047683] env[62460]: DEBUG oslo.service.loopingcall [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 957.050321] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 957.050744] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-68cad790-7e9a-4eb3-9c66-43bb8e2e65bb {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.073040] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 957.073040] env[62460]: value = "task-1313935" [ 957.073040] env[62460]: _type = "Task" [ 957.073040] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.080843] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313935, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.179373] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e23f31ca-eaa9-48d0-96ec-f3e8499490e7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.189598] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c436305a-7192-4b6b-a10c-1a18b646b87e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.222404] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61c56420-1a5e-4e9b-9662-54b749a73e86 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.230518] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89af185f-9654-4e67-8a62-5b1628397c39 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.246317] env[62460]: DEBUG nova.compute.provider_tree [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 957.308482] env[62460]: DEBUG nova.objects.instance [None req-6a39c0af-9dcd-4d60-85ea-cf89f8934ae2 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Lazy-loading 'flavor' on Instance uuid bf28fc98-d9a9-4678-a107-ace2b6503353 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 957.455667] env[62460]: DEBUG oslo_vmware.api [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313934, 'name': ReconfigVM_Task, 'duration_secs': 0.360331} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.455970] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281244', 'volume_id': 'c472b891-df79-428d-a905-5349e414e7f5', 'name': 'volume-c472b891-df79-428d-a905-5349e414e7f5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '896f8ff4-e45f-4403-a727-03ee25e58609', 'attached_at': '', 'detached_at': '', 'volume_id': 'c472b891-df79-428d-a905-5349e414e7f5', 'serial': 'c472b891-df79-428d-a905-5349e414e7f5'} {{(pid=62460) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 957.456564] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dc82ce07-a26f-45c0-8c60-779c0ccdd293 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.462053] env[62460]: DEBUG oslo_vmware.api [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 957.462053] env[62460]: value = "task-1313936" [ 957.462053] env[62460]: _type = "Task" [ 957.462053] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.469352] env[62460]: DEBUG oslo_vmware.api [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313936, 'name': Rename_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.584078] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313935, 'name': CreateVM_Task, 'duration_secs': 0.437192} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.584299] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 957.585069] env[62460]: DEBUG oslo_concurrency.lockutils [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.585278] env[62460]: DEBUG oslo_concurrency.lockutils [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.585649] env[62460]: DEBUG oslo_concurrency.lockutils [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 957.585937] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f67f9dae-58e2-42fc-96b2-2eeaae1481a8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.591265] env[62460]: DEBUG oslo_vmware.api [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 957.591265] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]5268ed61-87e0-68b7-045b-5e8b084cea7d" [ 957.591265] env[62460]: _type = "Task" [ 957.591265] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.600657] env[62460]: DEBUG oslo_vmware.api [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5268ed61-87e0-68b7-045b-5e8b084cea7d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.749399] env[62460]: DEBUG nova.scheduler.client.report [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 957.755032] env[62460]: DEBUG nova.compute.manager [req-50ca5a37-3afb-491a-9852-5b04b238953d req-f25c0272-5ddd-446c-a21e-eda95959b618 service nova] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Received event network-changed-9f9c5b7b-922a-4387-b997-3f552dd85a50 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 957.755292] env[62460]: DEBUG nova.compute.manager [req-50ca5a37-3afb-491a-9852-5b04b238953d req-f25c0272-5ddd-446c-a21e-eda95959b618 service nova] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Refreshing instance network info cache due to event network-changed-9f9c5b7b-922a-4387-b997-3f552dd85a50. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 957.755499] env[62460]: DEBUG oslo_concurrency.lockutils [req-50ca5a37-3afb-491a-9852-5b04b238953d req-f25c0272-5ddd-446c-a21e-eda95959b618 service nova] Acquiring lock "refresh_cache-290500b5-d5b4-43b6-a843-20dacf561f4b" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.755650] env[62460]: DEBUG oslo_concurrency.lockutils [req-50ca5a37-3afb-491a-9852-5b04b238953d req-f25c0272-5ddd-446c-a21e-eda95959b618 service nova] Acquired lock "refresh_cache-290500b5-d5b4-43b6-a843-20dacf561f4b" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.755817] env[62460]: DEBUG nova.network.neutron [req-50ca5a37-3afb-491a-9852-5b04b238953d req-f25c0272-5ddd-446c-a21e-eda95959b618 service nova] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Refreshing network info cache for port 9f9c5b7b-922a-4387-b997-3f552dd85a50 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 957.814085] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6a39c0af-9dcd-4d60-85ea-cf89f8934ae2 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Acquiring lock "refresh_cache-bf28fc98-d9a9-4678-a107-ace2b6503353" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.814330] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6a39c0af-9dcd-4d60-85ea-cf89f8934ae2 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Acquired lock "refresh_cache-bf28fc98-d9a9-4678-a107-ace2b6503353" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.973334] env[62460]: DEBUG oslo_vmware.api [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313936, 'name': Rename_Task, 'duration_secs': 0.269772} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.973673] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 957.973797] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b156def4-8574-4cf8-aa87-7ef8fa28e9cf {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.979859] env[62460]: DEBUG oslo_vmware.api [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 957.979859] env[62460]: value = "task-1313937" [ 957.979859] env[62460]: _type = "Task" [ 957.979859] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.987268] env[62460]: DEBUG oslo_vmware.api [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313937, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.000815] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0aa06aad-3d54-4fa6-8a41-6527ef18a645 tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Acquiring lock "003191d3-fe38-4bde-8d7c-75b4aad413b4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.001085] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0aa06aad-3d54-4fa6-8a41-6527ef18a645 tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Lock "003191d3-fe38-4bde-8d7c-75b4aad413b4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.001346] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0aa06aad-3d54-4fa6-8a41-6527ef18a645 tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Acquiring lock "003191d3-fe38-4bde-8d7c-75b4aad413b4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.001550] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0aa06aad-3d54-4fa6-8a41-6527ef18a645 tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Lock "003191d3-fe38-4bde-8d7c-75b4aad413b4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.001867] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0aa06aad-3d54-4fa6-8a41-6527ef18a645 tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Lock "003191d3-fe38-4bde-8d7c-75b4aad413b4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.003908] env[62460]: INFO nova.compute.manager [None req-0aa06aad-3d54-4fa6-8a41-6527ef18a645 tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Terminating instance [ 958.005731] env[62460]: DEBUG nova.compute.manager [None req-0aa06aad-3d54-4fa6-8a41-6527ef18a645 tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 958.006109] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-0aa06aad-3d54-4fa6-8a41-6527ef18a645 tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 958.006751] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22a543c1-49d2-490d-858d-998ca6a1eabc {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.013952] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-0aa06aad-3d54-4fa6-8a41-6527ef18a645 tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 958.014138] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-55b48081-05f8-42d4-9d3c-c342814e978a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.021542] env[62460]: DEBUG oslo_vmware.api [None req-0aa06aad-3d54-4fa6-8a41-6527ef18a645 tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Waiting for the task: (returnval){ [ 958.021542] env[62460]: value = "task-1313938" [ 958.021542] env[62460]: _type = "Task" [ 958.021542] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.033492] env[62460]: DEBUG oslo_vmware.api [None req-0aa06aad-3d54-4fa6-8a41-6527ef18a645 tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313938, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.067568] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0fa3495d-783c-4299-a9d5-dee125e75d0d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Acquiring lock "db331001-1744-4c42-a40c-87f3e60da4b8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.067879] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0fa3495d-783c-4299-a9d5-dee125e75d0d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Lock "db331001-1744-4c42-a40c-87f3e60da4b8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.068144] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0fa3495d-783c-4299-a9d5-dee125e75d0d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Acquiring lock "db331001-1744-4c42-a40c-87f3e60da4b8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.068478] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0fa3495d-783c-4299-a9d5-dee125e75d0d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Lock "db331001-1744-4c42-a40c-87f3e60da4b8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.068748] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0fa3495d-783c-4299-a9d5-dee125e75d0d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Lock "db331001-1744-4c42-a40c-87f3e60da4b8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.072090] env[62460]: INFO nova.compute.manager [None req-0fa3495d-783c-4299-a9d5-dee125e75d0d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Terminating instance [ 958.073248] env[62460]: DEBUG nova.compute.manager [None req-0fa3495d-783c-4299-a9d5-dee125e75d0d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 958.073467] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-0fa3495d-783c-4299-a9d5-dee125e75d0d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 958.074439] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63933375-ec86-46a2-8fd3-577028b55f30 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.082318] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fa3495d-783c-4299-a9d5-dee125e75d0d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 958.082585] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6d33800e-5dc9-4a9f-96fd-405143654bad {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.089731] env[62460]: DEBUG oslo_vmware.api [None req-0fa3495d-783c-4299-a9d5-dee125e75d0d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Waiting for the task: (returnval){ [ 958.089731] env[62460]: value = "task-1313939" [ 958.089731] env[62460]: _type = "Task" [ 958.089731] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.100543] env[62460]: DEBUG oslo_vmware.api [None req-0fa3495d-783c-4299-a9d5-dee125e75d0d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313939, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.104860] env[62460]: DEBUG oslo_vmware.api [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5268ed61-87e0-68b7-045b-5e8b084cea7d, 'name': SearchDatastore_Task, 'duration_secs': 0.015969} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.104860] env[62460]: DEBUG oslo_concurrency.lockutils [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.104860] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 958.105070] env[62460]: DEBUG oslo_concurrency.lockutils [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.105241] env[62460]: DEBUG oslo_concurrency.lockutils [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.105442] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 958.105719] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-37d8d316-5840-4a6b-96e5-446d701a0538 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.113474] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 958.113727] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 958.114541] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f9e0817e-89fa-40cc-8d0d-c62bce9ef797 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.119865] env[62460]: DEBUG oslo_vmware.api [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 958.119865] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]521c8f64-22de-5787-ec9c-5ac522a98001" [ 958.119865] env[62460]: _type = "Task" [ 958.119865] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.127166] env[62460]: DEBUG oslo_vmware.api [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]521c8f64-22de-5787-ec9c-5ac522a98001, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.170021] env[62460]: DEBUG nova.network.neutron [None req-6a39c0af-9dcd-4d60-85ea-cf89f8934ae2 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 958.258919] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.261s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.259557] env[62460]: DEBUG nova.compute.manager [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 958.264756] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.182s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.266481] env[62460]: INFO nova.compute.claims [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 958.468948] env[62460]: DEBUG nova.network.neutron [req-50ca5a37-3afb-491a-9852-5b04b238953d req-f25c0272-5ddd-446c-a21e-eda95959b618 service nova] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Updated VIF entry in instance network info cache for port 9f9c5b7b-922a-4387-b997-3f552dd85a50. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 958.469312] env[62460]: DEBUG nova.network.neutron [req-50ca5a37-3afb-491a-9852-5b04b238953d req-f25c0272-5ddd-446c-a21e-eda95959b618 service nova] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Updating instance_info_cache with network_info: [{"id": "9f9c5b7b-922a-4387-b997-3f552dd85a50", "address": "fa:16:3e:78:80:d5", "network": {"id": "de789887-9cf9-4614-a7e1-dab430aaabaf", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-656400207-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0da056d93bdf40c39d6e82e457727ff6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5116f690-f825-4fee-8a47-42b073e716c5", "external-id": "nsx-vlan-transportzone-692", "segmentation_id": 692, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f9c5b7b-92", "ovs_interfaceid": "9f9c5b7b-922a-4387-b997-3f552dd85a50", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.489137] env[62460]: DEBUG oslo_vmware.api [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1313937, 'name': PowerOnVM_Task, 'duration_secs': 0.503326} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.489401] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 958.531035] env[62460]: DEBUG oslo_vmware.api [None req-0aa06aad-3d54-4fa6-8a41-6527ef18a645 tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313938, 'name': PowerOffVM_Task, 'duration_secs': 0.185436} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.531319] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-0aa06aad-3d54-4fa6-8a41-6527ef18a645 tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 958.531501] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-0aa06aad-3d54-4fa6-8a41-6527ef18a645 tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 958.531757] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-29a60184-20c7-4367-8b35-0946a4b40f2f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.595446] env[62460]: DEBUG nova.compute.manager [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 958.596514] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd94ce64-df26-4cd5-b002-e987e2348bd8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.602916] env[62460]: DEBUG oslo_vmware.api [None req-0fa3495d-783c-4299-a9d5-dee125e75d0d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313939, 'name': PowerOffVM_Task, 'duration_secs': 0.218166} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.603632] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fa3495d-783c-4299-a9d5-dee125e75d0d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 958.603866] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-0fa3495d-783c-4299-a9d5-dee125e75d0d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 958.604150] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9b463546-c0a8-4142-accb-84f0f9ca8ec3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.630283] env[62460]: DEBUG oslo_vmware.api [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]521c8f64-22de-5787-ec9c-5ac522a98001, 'name': SearchDatastore_Task, 'duration_secs': 0.010806} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.632136] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-0aa06aad-3d54-4fa6-8a41-6527ef18a645 tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 958.632393] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-0aa06aad-3d54-4fa6-8a41-6527ef18a645 tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Deleting contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 958.632603] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-0aa06aad-3d54-4fa6-8a41-6527ef18a645 tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Deleting the datastore file [datastore1] 003191d3-fe38-4bde-8d7c-75b4aad413b4 {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 958.632856] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-81a51de1-97a7-419e-832b-4a3f459e579d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.635046] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2c151061-9e3b-458f-82b9-dcb4ad29c6e6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.639466] env[62460]: DEBUG oslo_vmware.api [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 958.639466] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52dfa52c-ce9e-a842-61bf-916e7e5360d5" [ 958.639466] env[62460]: _type = "Task" [ 958.639466] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.644969] env[62460]: DEBUG oslo_vmware.api [None req-0aa06aad-3d54-4fa6-8a41-6527ef18a645 tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Waiting for the task: (returnval){ [ 958.644969] env[62460]: value = "task-1313942" [ 958.644969] env[62460]: _type = "Task" [ 958.644969] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.651962] env[62460]: DEBUG oslo_vmware.api [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52dfa52c-ce9e-a842-61bf-916e7e5360d5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.657655] env[62460]: DEBUG oslo_vmware.api [None req-0aa06aad-3d54-4fa6-8a41-6527ef18a645 tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313942, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.685167] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-0fa3495d-783c-4299-a9d5-dee125e75d0d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 958.685427] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-0fa3495d-783c-4299-a9d5-dee125e75d0d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Deleting contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 958.685619] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-0fa3495d-783c-4299-a9d5-dee125e75d0d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Deleting the datastore file [datastore1] db331001-1744-4c42-a40c-87f3e60da4b8 {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 958.686107] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5a0e1835-c0f1-4761-b122-add3ddef3e28 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.695035] env[62460]: DEBUG oslo_vmware.api [None req-0fa3495d-783c-4299-a9d5-dee125e75d0d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Waiting for the task: (returnval){ [ 958.695035] env[62460]: value = "task-1313943" [ 958.695035] env[62460]: _type = "Task" [ 958.695035] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.703523] env[62460]: DEBUG oslo_vmware.api [None req-0fa3495d-783c-4299-a9d5-dee125e75d0d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313943, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.766471] env[62460]: DEBUG nova.compute.utils [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 958.768150] env[62460]: DEBUG nova.compute.manager [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 958.768363] env[62460]: DEBUG nova.network.neutron [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 958.839229] env[62460]: DEBUG nova.policy [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '632d7ca8cb4243f990d741a3f70f422a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '10ac54db1ef54e249a077bbb3dde4242', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 958.972498] env[62460]: DEBUG oslo_concurrency.lockutils [req-50ca5a37-3afb-491a-9852-5b04b238953d req-f25c0272-5ddd-446c-a21e-eda95959b618 service nova] Releasing lock "refresh_cache-290500b5-d5b4-43b6-a843-20dacf561f4b" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.996076] env[62460]: DEBUG nova.network.neutron [None req-6a39c0af-9dcd-4d60-85ea-cf89f8934ae2 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Updating instance_info_cache with network_info: [{"id": "7bb00d79-430f-44e2-9e92-57109c05d0bb", "address": "fa:16:3e:94:48:0c", "network": {"id": "e2ad0587-12a2-40ce-b2d0-c682042335fb", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1549110976-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.202", "type": "floating", "version": 4, "meta": {}}]}, {"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1845e17fb7e64cf594c4e81c67ad8243", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c405e9f-a6c8-4308-acac-071654efe18e", "external-id": "nsx-vlan-transportzone-851", "segmentation_id": 851, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7bb00d79-43", "ovs_interfaceid": "7bb00d79-430f-44e2-9e92-57109c05d0bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 959.119375] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2a2629a6-83db-4df5-b0d1-7eeca34e4e74 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lock "896f8ff4-e45f-4403-a727-03ee25e58609" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 36.863s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.152719] env[62460]: DEBUG oslo_vmware.api [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52dfa52c-ce9e-a842-61bf-916e7e5360d5, 'name': SearchDatastore_Task, 'duration_secs': 0.013411} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.153614] env[62460]: DEBUG nova.network.neutron [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Successfully created port: 352f5563-2248-439c-a843-2001a2d5d230 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 959.155676] env[62460]: DEBUG oslo_concurrency.lockutils [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.155951] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 290500b5-d5b4-43b6-a843-20dacf561f4b/290500b5-d5b4-43b6-a843-20dacf561f4b.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 959.156802] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-91845a16-57a8-4d22-bb27-64b59af670d4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.161064] env[62460]: DEBUG oslo_vmware.api [None req-0aa06aad-3d54-4fa6-8a41-6527ef18a645 tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313942, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.236502} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.161660] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-0aa06aad-3d54-4fa6-8a41-6527ef18a645 tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 959.161868] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-0aa06aad-3d54-4fa6-8a41-6527ef18a645 tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Deleted contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 959.162074] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-0aa06aad-3d54-4fa6-8a41-6527ef18a645 tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 959.162265] env[62460]: INFO nova.compute.manager [None req-0aa06aad-3d54-4fa6-8a41-6527ef18a645 tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Took 1.16 seconds to destroy the instance on the hypervisor. [ 959.162511] env[62460]: DEBUG oslo.service.loopingcall [None req-0aa06aad-3d54-4fa6-8a41-6527ef18a645 tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 959.162710] env[62460]: DEBUG nova.compute.manager [-] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 959.162805] env[62460]: DEBUG nova.network.neutron [-] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 959.165929] env[62460]: DEBUG oslo_vmware.api [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 959.165929] env[62460]: value = "task-1313944" [ 959.165929] env[62460]: _type = "Task" [ 959.165929] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.174073] env[62460]: DEBUG oslo_vmware.api [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313944, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.204815] env[62460]: DEBUG oslo_vmware.api [None req-0fa3495d-783c-4299-a9d5-dee125e75d0d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Task: {'id': task-1313943, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.265005} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.205088] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-0fa3495d-783c-4299-a9d5-dee125e75d0d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 959.205357] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-0fa3495d-783c-4299-a9d5-dee125e75d0d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Deleted contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 959.205564] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-0fa3495d-783c-4299-a9d5-dee125e75d0d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 959.205746] env[62460]: INFO nova.compute.manager [None req-0fa3495d-783c-4299-a9d5-dee125e75d0d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Took 1.13 seconds to destroy the instance on the hypervisor. [ 959.205996] env[62460]: DEBUG oslo.service.loopingcall [None req-0fa3495d-783c-4299-a9d5-dee125e75d0d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 959.206971] env[62460]: DEBUG nova.compute.manager [-] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 959.206971] env[62460]: DEBUG nova.network.neutron [-] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 959.278027] env[62460]: DEBUG nova.compute.manager [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 959.453247] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1de5df41-5d11-43c8-9708-d4900076226b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "interface-abd4501b-3ad8-4103-921d-5b80fc2f313f-50d6a0c4-3681-4c1a-b42f-2987ee7a3bfb" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.454298] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1de5df41-5d11-43c8-9708-d4900076226b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "interface-abd4501b-3ad8-4103-921d-5b80fc2f313f-50d6a0c4-3681-4c1a-b42f-2987ee7a3bfb" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.454298] env[62460]: DEBUG nova.objects.instance [None req-1de5df41-5d11-43c8-9708-d4900076226b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lazy-loading 'flavor' on Instance uuid abd4501b-3ad8-4103-921d-5b80fc2f313f {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 959.457568] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9665b06b-2963-4143-80ab-42a5c799df54 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.465649] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ab2b2ea-14ba-43c6-89fa-ad73d97a3d3d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.499582] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6a39c0af-9dcd-4d60-85ea-cf89f8934ae2 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Releasing lock "refresh_cache-bf28fc98-d9a9-4678-a107-ace2b6503353" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.499887] env[62460]: DEBUG nova.compute.manager [None req-6a39c0af-9dcd-4d60-85ea-cf89f8934ae2 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Inject network info {{(pid=62460) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7219}} [ 959.500173] env[62460]: DEBUG nova.compute.manager [None req-6a39c0af-9dcd-4d60-85ea-cf89f8934ae2 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] network_info to inject: |[{"id": "7bb00d79-430f-44e2-9e92-57109c05d0bb", "address": "fa:16:3e:94:48:0c", "network": {"id": "e2ad0587-12a2-40ce-b2d0-c682042335fb", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1549110976-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.202", "type": "floating", "version": 4, "meta": {}}]}, {"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1845e17fb7e64cf594c4e81c67ad8243", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c405e9f-a6c8-4308-acac-071654efe18e", "external-id": "nsx-vlan-transportzone-851", "segmentation_id": 851, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7bb00d79-43", "ovs_interfaceid": "7bb00d79-430f-44e2-9e92-57109c05d0bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7220}} [ 959.505273] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-6a39c0af-9dcd-4d60-85ea-cf89f8934ae2 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Reconfiguring VM instance to set the machine id {{(pid=62460) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1796}} [ 959.506181] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0bafb713-2fb8-4b08-aecd-9d8f202fc9e2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.518151] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-145a6d40-3b3b-4647-8ae4-e68b48495c15 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.529009] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-232b8d54-dcd1-4900-bf8f-3612040c61a9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.533489] env[62460]: DEBUG oslo_vmware.api [None req-6a39c0af-9dcd-4d60-85ea-cf89f8934ae2 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Waiting for the task: (returnval){ [ 959.533489] env[62460]: value = "task-1313945" [ 959.533489] env[62460]: _type = "Task" [ 959.533489] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.546143] env[62460]: DEBUG nova.compute.provider_tree [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 959.554802] env[62460]: DEBUG oslo_vmware.api [None req-6a39c0af-9dcd-4d60-85ea-cf89f8934ae2 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Task: {'id': task-1313945, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.681167] env[62460]: DEBUG oslo_vmware.api [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313944, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.845487] env[62460]: DEBUG nova.compute.manager [req-781e470b-d257-4d80-940e-605e954f3b7a req-a9beb488-fa2a-415b-b4a2-3b6e181d8add service nova] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Received event network-vif-deleted-96b72a46-8d24-420d-a0f2-407651530be5 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 959.845746] env[62460]: INFO nova.compute.manager [req-781e470b-d257-4d80-940e-605e954f3b7a req-a9beb488-fa2a-415b-b4a2-3b6e181d8add service nova] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Neutron deleted interface 96b72a46-8d24-420d-a0f2-407651530be5; detaching it from the instance and deleting it from the info cache [ 959.845882] env[62460]: DEBUG nova.network.neutron [req-781e470b-d257-4d80-940e-605e954f3b7a req-a9beb488-fa2a-415b-b4a2-3b6e181d8add service nova] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 959.871120] env[62460]: DEBUG nova.objects.instance [None req-e02d4b37-ae77-44da-acd2-da4ff1500a31 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Lazy-loading 'flavor' on Instance uuid bf28fc98-d9a9-4678-a107-ace2b6503353 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 959.890158] env[62460]: DEBUG nova.compute.manager [req-25c00d3f-e74c-4162-a144-7f83c6a0b2a7 req-12a49752-9556-4d4f-96bd-2e58d9152c4b service nova] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Received event network-changed-7bb00d79-430f-44e2-9e92-57109c05d0bb {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 959.890158] env[62460]: DEBUG nova.compute.manager [req-25c00d3f-e74c-4162-a144-7f83c6a0b2a7 req-12a49752-9556-4d4f-96bd-2e58d9152c4b service nova] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Refreshing instance network info cache due to event network-changed-7bb00d79-430f-44e2-9e92-57109c05d0bb. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 959.890158] env[62460]: DEBUG oslo_concurrency.lockutils [req-25c00d3f-e74c-4162-a144-7f83c6a0b2a7 req-12a49752-9556-4d4f-96bd-2e58d9152c4b service nova] Acquiring lock "refresh_cache-bf28fc98-d9a9-4678-a107-ace2b6503353" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.890158] env[62460]: DEBUG oslo_concurrency.lockutils [req-25c00d3f-e74c-4162-a144-7f83c6a0b2a7 req-12a49752-9556-4d4f-96bd-2e58d9152c4b service nova] Acquired lock "refresh_cache-bf28fc98-d9a9-4678-a107-ace2b6503353" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.890572] env[62460]: DEBUG nova.network.neutron [req-25c00d3f-e74c-4162-a144-7f83c6a0b2a7 req-12a49752-9556-4d4f-96bd-2e58d9152c4b service nova] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Refreshing network info cache for port 7bb00d79-430f-44e2-9e92-57109c05d0bb {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 959.906289] env[62460]: DEBUG nova.network.neutron [-] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 960.047144] env[62460]: DEBUG oslo_vmware.api [None req-6a39c0af-9dcd-4d60-85ea-cf89f8934ae2 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Task: {'id': task-1313945, 'name': ReconfigVM_Task, 'duration_secs': 0.156059} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.047523] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-6a39c0af-9dcd-4d60-85ea-cf89f8934ae2 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Reconfigured VM instance to set the machine id {{(pid=62460) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1799}} [ 960.063393] env[62460]: DEBUG nova.objects.instance [None req-1de5df41-5d11-43c8-9708-d4900076226b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lazy-loading 'pci_requests' on Instance uuid abd4501b-3ad8-4103-921d-5b80fc2f313f {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 960.084106] env[62460]: DEBUG nova.scheduler.client.report [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Updated inventory for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with generation 119 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 960.084454] env[62460]: DEBUG nova.compute.provider_tree [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Updating resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 generation from 119 to 120 during operation: update_inventory {{(pid=62460) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 960.084564] env[62460]: DEBUG nova.compute.provider_tree [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 960.122538] env[62460]: DEBUG nova.network.neutron [-] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 960.176446] env[62460]: DEBUG oslo_vmware.api [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313944, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.777688} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.176727] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 290500b5-d5b4-43b6-a843-20dacf561f4b/290500b5-d5b4-43b6-a843-20dacf561f4b.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 960.176951] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 960.177225] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9f41a2a7-097a-49be-90d3-08b311d727e8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.183502] env[62460]: DEBUG oslo_vmware.api [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 960.183502] env[62460]: value = "task-1313946" [ 960.183502] env[62460]: _type = "Task" [ 960.183502] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.190711] env[62460]: DEBUG oslo_vmware.api [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313946, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.290136] env[62460]: DEBUG nova.compute.manager [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 960.314325] env[62460]: DEBUG nova.virt.hardware [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 960.314594] env[62460]: DEBUG nova.virt.hardware [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 960.314758] env[62460]: DEBUG nova.virt.hardware [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 960.314948] env[62460]: DEBUG nova.virt.hardware [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 960.315116] env[62460]: DEBUG nova.virt.hardware [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 960.315274] env[62460]: DEBUG nova.virt.hardware [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 960.315483] env[62460]: DEBUG nova.virt.hardware [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 960.315646] env[62460]: DEBUG nova.virt.hardware [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 960.315816] env[62460]: DEBUG nova.virt.hardware [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 960.316009] env[62460]: DEBUG nova.virt.hardware [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 960.316211] env[62460]: DEBUG nova.virt.hardware [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 960.317064] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7214e2f9-2690-4b73-8cee-55dfad3d1fd0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.324860] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5946cecb-40e2-4bae-ad1c-f8ef1fccf874 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.348381] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0d5ca7cd-bd7a-40e8-8f12-1630b1f97495 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.356173] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13a83333-bd72-4a2a-affe-c9ae96328f09 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.380980] env[62460]: DEBUG nova.compute.manager [req-781e470b-d257-4d80-940e-605e954f3b7a req-a9beb488-fa2a-415b-b4a2-3b6e181d8add service nova] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Detach interface failed, port_id=96b72a46-8d24-420d-a0f2-407651530be5, reason: Instance db331001-1744-4c42-a40c-87f3e60da4b8 could not be found. {{(pid=62460) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 960.381574] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e02d4b37-ae77-44da-acd2-da4ff1500a31 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Acquiring lock "refresh_cache-bf28fc98-d9a9-4678-a107-ace2b6503353" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.408971] env[62460]: INFO nova.compute.manager [-] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Took 1.25 seconds to deallocate network for instance. [ 960.568855] env[62460]: DEBUG nova.objects.base [None req-1de5df41-5d11-43c8-9708-d4900076226b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=62460) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 960.569148] env[62460]: DEBUG nova.network.neutron [None req-1de5df41-5d11-43c8-9708-d4900076226b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 960.589232] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.324s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.589719] env[62460]: DEBUG nova.compute.manager [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 960.592125] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d1afac0f-2ef1-496a-b856-f36557e80c73 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.123s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.593233] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d1afac0f-2ef1-496a-b856-f36557e80c73 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.618784] env[62460]: INFO nova.scheduler.client.report [None req-d1afac0f-2ef1-496a-b856-f36557e80c73 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Deleted allocations for instance 6b084ed4-e8f0-4063-bf2a-252740640753 [ 960.620597] env[62460]: DEBUG nova.network.neutron [req-25c00d3f-e74c-4162-a144-7f83c6a0b2a7 req-12a49752-9556-4d4f-96bd-2e58d9152c4b service nova] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Updated VIF entry in instance network info cache for port 7bb00d79-430f-44e2-9e92-57109c05d0bb. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 960.620759] env[62460]: DEBUG nova.network.neutron [req-25c00d3f-e74c-4162-a144-7f83c6a0b2a7 req-12a49752-9556-4d4f-96bd-2e58d9152c4b service nova] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Updating instance_info_cache with network_info: [{"id": "7bb00d79-430f-44e2-9e92-57109c05d0bb", "address": "fa:16:3e:94:48:0c", "network": {"id": "e2ad0587-12a2-40ce-b2d0-c682042335fb", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1549110976-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.202", "type": "floating", "version": 4, "meta": {}}]}, {"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1845e17fb7e64cf594c4e81c67ad8243", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c405e9f-a6c8-4308-acac-071654efe18e", "external-id": "nsx-vlan-transportzone-851", "segmentation_id": 851, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7bb00d79-43", "ovs_interfaceid": "7bb00d79-430f-44e2-9e92-57109c05d0bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 960.624335] env[62460]: INFO nova.compute.manager [-] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Took 1.42 seconds to deallocate network for instance. [ 960.645154] env[62460]: DEBUG nova.policy [None req-1de5df41-5d11-43c8-9708-d4900076226b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd7b38fb0a53f4d078ae86d68e3eb6268', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '53fa06008e9f43488362895e7a143700', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 960.693427] env[62460]: DEBUG oslo_vmware.api [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313946, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068098} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.693542] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 960.694250] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a68850a2-4e5a-4562-87c8-81ef54b8c9c5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.724960] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Reconfiguring VM instance instance-0000005e to attach disk [datastore1] 290500b5-d5b4-43b6-a843-20dacf561f4b/290500b5-d5b4-43b6-a843-20dacf561f4b.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 960.725303] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f13bedb0-21b2-44ec-b84b-fc104a4067cc {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.744328] env[62460]: DEBUG oslo_vmware.api [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 960.744328] env[62460]: value = "task-1313947" [ 960.744328] env[62460]: _type = "Task" [ 960.744328] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.755221] env[62460]: DEBUG oslo_vmware.api [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313947, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.915205] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0aa06aad-3d54-4fa6-8a41-6527ef18a645 tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.915512] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0aa06aad-3d54-4fa6-8a41-6527ef18a645 tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.915745] env[62460]: DEBUG nova.objects.instance [None req-0aa06aad-3d54-4fa6-8a41-6527ef18a645 tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Lazy-loading 'resources' on Instance uuid 003191d3-fe38-4bde-8d7c-75b4aad413b4 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 960.961546] env[62460]: DEBUG nova.network.neutron [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Successfully updated port: 352f5563-2248-439c-a843-2001a2d5d230 {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 961.096317] env[62460]: DEBUG nova.compute.utils [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 961.098007] env[62460]: DEBUG nova.compute.manager [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 961.098246] env[62460]: DEBUG nova.network.neutron [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 961.125034] env[62460]: DEBUG oslo_concurrency.lockutils [req-25c00d3f-e74c-4162-a144-7f83c6a0b2a7 req-12a49752-9556-4d4f-96bd-2e58d9152c4b service nova] Releasing lock "refresh_cache-bf28fc98-d9a9-4678-a107-ace2b6503353" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.125379] env[62460]: DEBUG nova.compute.manager [req-25c00d3f-e74c-4162-a144-7f83c6a0b2a7 req-12a49752-9556-4d4f-96bd-2e58d9152c4b service nova] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Received event network-vif-deleted-73df6556-a3c8-4ee6-9881-8b978f97ca6f {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 961.125606] env[62460]: INFO nova.compute.manager [req-25c00d3f-e74c-4162-a144-7f83c6a0b2a7 req-12a49752-9556-4d4f-96bd-2e58d9152c4b service nova] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Neutron deleted interface 73df6556-a3c8-4ee6-9881-8b978f97ca6f; detaching it from the instance and deleting it from the info cache [ 961.125819] env[62460]: DEBUG nova.network.neutron [req-25c00d3f-e74c-4162-a144-7f83c6a0b2a7 req-12a49752-9556-4d4f-96bd-2e58d9152c4b service nova] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.130785] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e02d4b37-ae77-44da-acd2-da4ff1500a31 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Acquired lock "refresh_cache-bf28fc98-d9a9-4678-a107-ace2b6503353" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.132452] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d1afac0f-2ef1-496a-b856-f36557e80c73 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "6b084ed4-e8f0-4063-bf2a-252740640753" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.659s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.134632] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0fa3495d-783c-4299-a9d5-dee125e75d0d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.136801] env[62460]: DEBUG nova.policy [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e531b7c3e7544e53b126d1336ac44c99', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bfa5a65b0f614e769de5b3aa77bf869e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 961.256599] env[62460]: DEBUG oslo_vmware.api [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313947, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.464705] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "refresh_cache-f0f66557-04ef-4a96-8a56-e640bfda3b36" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 961.464853] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquired lock "refresh_cache-f0f66557-04ef-4a96-8a56-e640bfda3b36" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.465019] env[62460]: DEBUG nova.network.neutron [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 961.567823] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe8b7cf9-7bcd-4653-ba1f-ee85fdcb663a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.575922] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1225e94-fbb6-4965-9e25-55d25a1ae639 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.617264] env[62460]: DEBUG nova.compute.manager [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 961.625018] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c249c615-c68b-48fe-87da-b3166a6bf0f9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.630980] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd3b66bf-16d4-4b56-94a9-aa92e11c1b04 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.635973] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d155d9c8-f236-46f3-8a98-010568346e1e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.648215] env[62460]: DEBUG nova.compute.provider_tree [None req-0aa06aad-3d54-4fa6-8a41-6527ef18a645 tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 961.653699] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf424a5e-4c32-46ac-b4d8-1061576ede73 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.681578] env[62460]: DEBUG nova.compute.manager [req-25c00d3f-e74c-4162-a144-7f83c6a0b2a7 req-12a49752-9556-4d4f-96bd-2e58d9152c4b service nova] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Detach interface failed, port_id=73df6556-a3c8-4ee6-9881-8b978f97ca6f, reason: Instance 003191d3-fe38-4bde-8d7c-75b4aad413b4 could not be found. {{(pid=62460) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 961.756021] env[62460]: DEBUG oslo_vmware.api [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313947, 'name': ReconfigVM_Task, 'duration_secs': 0.519752} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.756385] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Reconfigured VM instance instance-0000005e to attach disk [datastore1] 290500b5-d5b4-43b6-a843-20dacf561f4b/290500b5-d5b4-43b6-a843-20dacf561f4b.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 961.757297] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-56970c4e-9f5d-45d2-8da8-cce790300fdf {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.763326] env[62460]: DEBUG oslo_vmware.api [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 961.763326] env[62460]: value = "task-1313948" [ 961.763326] env[62460]: _type = "Task" [ 961.763326] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.771961] env[62460]: DEBUG oslo_vmware.api [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313948, 'name': Rename_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.825593] env[62460]: DEBUG nova.network.neutron [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Successfully created port: 786ef177-843f-44d0-8920-2f4332e29154 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 961.877058] env[62460]: DEBUG nova.compute.manager [req-b855487b-cf08-43f5-bbe3-d7feb8b295ba req-5124bc5c-fe94-467c-b6b8-80a216d82abf service nova] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Received event network-vif-plugged-352f5563-2248-439c-a843-2001a2d5d230 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 961.877333] env[62460]: DEBUG oslo_concurrency.lockutils [req-b855487b-cf08-43f5-bbe3-d7feb8b295ba req-5124bc5c-fe94-467c-b6b8-80a216d82abf service nova] Acquiring lock "f0f66557-04ef-4a96-8a56-e640bfda3b36-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.877553] env[62460]: DEBUG oslo_concurrency.lockutils [req-b855487b-cf08-43f5-bbe3-d7feb8b295ba req-5124bc5c-fe94-467c-b6b8-80a216d82abf service nova] Lock "f0f66557-04ef-4a96-8a56-e640bfda3b36-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.877757] env[62460]: DEBUG oslo_concurrency.lockutils [req-b855487b-cf08-43f5-bbe3-d7feb8b295ba req-5124bc5c-fe94-467c-b6b8-80a216d82abf service nova] Lock "f0f66557-04ef-4a96-8a56-e640bfda3b36-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.877885] env[62460]: DEBUG nova.compute.manager [req-b855487b-cf08-43f5-bbe3-d7feb8b295ba req-5124bc5c-fe94-467c-b6b8-80a216d82abf service nova] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] No waiting events found dispatching network-vif-plugged-352f5563-2248-439c-a843-2001a2d5d230 {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 961.878364] env[62460]: WARNING nova.compute.manager [req-b855487b-cf08-43f5-bbe3-d7feb8b295ba req-5124bc5c-fe94-467c-b6b8-80a216d82abf service nova] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Received unexpected event network-vif-plugged-352f5563-2248-439c-a843-2001a2d5d230 for instance with vm_state building and task_state spawning. [ 961.878639] env[62460]: DEBUG nova.compute.manager [req-b855487b-cf08-43f5-bbe3-d7feb8b295ba req-5124bc5c-fe94-467c-b6b8-80a216d82abf service nova] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Received event network-changed-352f5563-2248-439c-a843-2001a2d5d230 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 961.878878] env[62460]: DEBUG nova.compute.manager [req-b855487b-cf08-43f5-bbe3-d7feb8b295ba req-5124bc5c-fe94-467c-b6b8-80a216d82abf service nova] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Refreshing instance network info cache due to event network-changed-352f5563-2248-439c-a843-2001a2d5d230. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 961.879404] env[62460]: DEBUG oslo_concurrency.lockutils [req-b855487b-cf08-43f5-bbe3-d7feb8b295ba req-5124bc5c-fe94-467c-b6b8-80a216d82abf service nova] Acquiring lock "refresh_cache-f0f66557-04ef-4a96-8a56-e640bfda3b36" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 961.880140] env[62460]: DEBUG nova.network.neutron [None req-e02d4b37-ae77-44da-acd2-da4ff1500a31 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 962.032776] env[62460]: DEBUG nova.network.neutron [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 962.119115] env[62460]: DEBUG nova.network.neutron [None req-1de5df41-5d11-43c8-9708-d4900076226b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Successfully updated port: 50d6a0c4-3681-4c1a-b42f-2987ee7a3bfb {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 962.143165] env[62460]: DEBUG nova.compute.manager [req-d1784f4f-2fdd-48b1-b2b1-f546b7e5c2ee req-ee7139e4-ae20-4133-a402-bf458cc94857 service nova] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Received event network-vif-plugged-50d6a0c4-3681-4c1a-b42f-2987ee7a3bfb {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 962.143165] env[62460]: DEBUG oslo_concurrency.lockutils [req-d1784f4f-2fdd-48b1-b2b1-f546b7e5c2ee req-ee7139e4-ae20-4133-a402-bf458cc94857 service nova] Acquiring lock "abd4501b-3ad8-4103-921d-5b80fc2f313f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.143541] env[62460]: DEBUG oslo_concurrency.lockutils [req-d1784f4f-2fdd-48b1-b2b1-f546b7e5c2ee req-ee7139e4-ae20-4133-a402-bf458cc94857 service nova] Lock "abd4501b-3ad8-4103-921d-5b80fc2f313f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.143541] env[62460]: DEBUG oslo_concurrency.lockutils [req-d1784f4f-2fdd-48b1-b2b1-f546b7e5c2ee req-ee7139e4-ae20-4133-a402-bf458cc94857 service nova] Lock "abd4501b-3ad8-4103-921d-5b80fc2f313f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.143696] env[62460]: DEBUG nova.compute.manager [req-d1784f4f-2fdd-48b1-b2b1-f546b7e5c2ee req-ee7139e4-ae20-4133-a402-bf458cc94857 service nova] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] No waiting events found dispatching network-vif-plugged-50d6a0c4-3681-4c1a-b42f-2987ee7a3bfb {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 962.143864] env[62460]: WARNING nova.compute.manager [req-d1784f4f-2fdd-48b1-b2b1-f546b7e5c2ee req-ee7139e4-ae20-4133-a402-bf458cc94857 service nova] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Received unexpected event network-vif-plugged-50d6a0c4-3681-4c1a-b42f-2987ee7a3bfb for instance with vm_state active and task_state None. [ 962.195664] env[62460]: DEBUG nova.scheduler.client.report [None req-0aa06aad-3d54-4fa6-8a41-6527ef18a645 tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Updated inventory for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with generation 120 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 962.196080] env[62460]: DEBUG nova.compute.provider_tree [None req-0aa06aad-3d54-4fa6-8a41-6527ef18a645 tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Updating resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 generation from 120 to 121 during operation: update_inventory {{(pid=62460) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 962.196364] env[62460]: DEBUG nova.compute.provider_tree [None req-0aa06aad-3d54-4fa6-8a41-6527ef18a645 tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 962.274282] env[62460]: DEBUG oslo_vmware.api [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313948, 'name': Rename_Task, 'duration_secs': 0.425508} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.274568] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 962.275085] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-90b7276a-4ffd-4425-bd68-e25e00d1af13 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.283017] env[62460]: DEBUG oslo_vmware.api [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 962.283017] env[62460]: value = "task-1313949" [ 962.283017] env[62460]: _type = "Task" [ 962.283017] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.286128] env[62460]: DEBUG nova.network.neutron [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Updating instance_info_cache with network_info: [{"id": "352f5563-2248-439c-a843-2001a2d5d230", "address": "fa:16:3e:fd:52:b2", "network": {"id": "c8b3d824-c8f0-4f7b-a91f-1987e04a5f09", "bridge": "br-int", "label": "tempest-ServersTestJSON-1769082812-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10ac54db1ef54e249a077bbb3dde4242", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "255460d5-71d4-4bfd-87f1-acc10085db7f", "external-id": "nsx-vlan-transportzone-152", "segmentation_id": 152, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap352f5563-22", "ovs_interfaceid": "352f5563-2248-439c-a843-2001a2d5d230", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.292057] env[62460]: DEBUG oslo_vmware.api [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313949, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.621848] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1de5df41-5d11-43c8-9708-d4900076226b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "refresh_cache-abd4501b-3ad8-4103-921d-5b80fc2f313f" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.622065] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1de5df41-5d11-43c8-9708-d4900076226b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquired lock "refresh_cache-abd4501b-3ad8-4103-921d-5b80fc2f313f" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.622255] env[62460]: DEBUG nova.network.neutron [None req-1de5df41-5d11-43c8-9708-d4900076226b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 962.627930] env[62460]: DEBUG nova.compute.manager [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 962.644207] env[62460]: DEBUG nova.network.neutron [None req-e02d4b37-ae77-44da-acd2-da4ff1500a31 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Updating instance_info_cache with network_info: [{"id": "7bb00d79-430f-44e2-9e92-57109c05d0bb", "address": "fa:16:3e:94:48:0c", "network": {"id": "e2ad0587-12a2-40ce-b2d0-c682042335fb", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1549110976-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.202", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1845e17fb7e64cf594c4e81c67ad8243", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c405e9f-a6c8-4308-acac-071654efe18e", "external-id": "nsx-vlan-transportzone-851", "segmentation_id": 851, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7bb00d79-43", "ovs_interfaceid": "7bb00d79-430f-44e2-9e92-57109c05d0bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.653182] env[62460]: DEBUG nova.virt.hardware [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 962.653438] env[62460]: DEBUG nova.virt.hardware [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 962.653613] env[62460]: DEBUG nova.virt.hardware [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 962.654608] env[62460]: DEBUG nova.virt.hardware [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 962.654608] env[62460]: DEBUG nova.virt.hardware [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 962.654608] env[62460]: DEBUG nova.virt.hardware [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 962.654608] env[62460]: DEBUG nova.virt.hardware [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 962.654608] env[62460]: DEBUG nova.virt.hardware [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 962.654820] env[62460]: DEBUG nova.virt.hardware [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 962.654893] env[62460]: DEBUG nova.virt.hardware [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 962.655086] env[62460]: DEBUG nova.virt.hardware [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 962.656035] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-605ea427-4dfb-47e3-97a8-1b07606b1fff {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.665095] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10f317f7-4334-41a2-8519-f877808a6ce1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.700699] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0aa06aad-3d54-4fa6-8a41-6527ef18a645 tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.785s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.702766] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0fa3495d-783c-4299-a9d5-dee125e75d0d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.568s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.703010] env[62460]: DEBUG nova.objects.instance [None req-0fa3495d-783c-4299-a9d5-dee125e75d0d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Lazy-loading 'resources' on Instance uuid db331001-1744-4c42-a40c-87f3e60da4b8 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 962.724658] env[62460]: INFO nova.scheduler.client.report [None req-0aa06aad-3d54-4fa6-8a41-6527ef18a645 tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Deleted allocations for instance 003191d3-fe38-4bde-8d7c-75b4aad413b4 [ 962.789435] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Releasing lock "refresh_cache-f0f66557-04ef-4a96-8a56-e640bfda3b36" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.790121] env[62460]: DEBUG nova.compute.manager [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Instance network_info: |[{"id": "352f5563-2248-439c-a843-2001a2d5d230", "address": "fa:16:3e:fd:52:b2", "network": {"id": "c8b3d824-c8f0-4f7b-a91f-1987e04a5f09", "bridge": "br-int", "label": "tempest-ServersTestJSON-1769082812-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10ac54db1ef54e249a077bbb3dde4242", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "255460d5-71d4-4bfd-87f1-acc10085db7f", "external-id": "nsx-vlan-transportzone-152", "segmentation_id": 152, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap352f5563-22", "ovs_interfaceid": "352f5563-2248-439c-a843-2001a2d5d230", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 962.794237] env[62460]: DEBUG oslo_concurrency.lockutils [req-b855487b-cf08-43f5-bbe3-d7feb8b295ba req-5124bc5c-fe94-467c-b6b8-80a216d82abf service nova] Acquired lock "refresh_cache-f0f66557-04ef-4a96-8a56-e640bfda3b36" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.794480] env[62460]: DEBUG nova.network.neutron [req-b855487b-cf08-43f5-bbe3-d7feb8b295ba req-5124bc5c-fe94-467c-b6b8-80a216d82abf service nova] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Refreshing network info cache for port 352f5563-2248-439c-a843-2001a2d5d230 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 962.795728] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fd:52:b2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '255460d5-71d4-4bfd-87f1-acc10085db7f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '352f5563-2248-439c-a843-2001a2d5d230', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 962.804189] env[62460]: DEBUG oslo.service.loopingcall [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 962.804362] env[62460]: DEBUG oslo_vmware.api [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313949, 'name': PowerOnVM_Task} progress is 99%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.805361] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 962.805610] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7e7606ef-062c-4cf3-a9de-f807574cbaaa {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.825275] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 962.825275] env[62460]: value = "task-1313950" [ 962.825275] env[62460]: _type = "Task" [ 962.825275] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.835581] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313950, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.148759] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e02d4b37-ae77-44da-acd2-da4ff1500a31 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Releasing lock "refresh_cache-bf28fc98-d9a9-4678-a107-ace2b6503353" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.149115] env[62460]: DEBUG nova.compute.manager [None req-e02d4b37-ae77-44da-acd2-da4ff1500a31 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Inject network info {{(pid=62460) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7219}} [ 963.149383] env[62460]: DEBUG nova.compute.manager [None req-e02d4b37-ae77-44da-acd2-da4ff1500a31 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] network_info to inject: |[{"id": "7bb00d79-430f-44e2-9e92-57109c05d0bb", "address": "fa:16:3e:94:48:0c", "network": {"id": "e2ad0587-12a2-40ce-b2d0-c682042335fb", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1549110976-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.202", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1845e17fb7e64cf594c4e81c67ad8243", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c405e9f-a6c8-4308-acac-071654efe18e", "external-id": "nsx-vlan-transportzone-851", "segmentation_id": 851, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7bb00d79-43", "ovs_interfaceid": "7bb00d79-430f-44e2-9e92-57109c05d0bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7220}} [ 963.154404] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-e02d4b37-ae77-44da-acd2-da4ff1500a31 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Reconfiguring VM instance to set the machine id {{(pid=62460) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1796}} [ 963.154994] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7c974481-6962-4fdb-b685-a003327419a7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.165532] env[62460]: WARNING nova.network.neutron [None req-1de5df41-5d11-43c8-9708-d4900076226b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] d3136e32-ad55-4b73-835c-8fa5f0480767 already exists in list: networks containing: ['d3136e32-ad55-4b73-835c-8fa5f0480767']. ignoring it [ 963.172689] env[62460]: DEBUG oslo_vmware.api [None req-e02d4b37-ae77-44da-acd2-da4ff1500a31 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Waiting for the task: (returnval){ [ 963.172689] env[62460]: value = "task-1313951" [ 963.172689] env[62460]: _type = "Task" [ 963.172689] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.182781] env[62460]: DEBUG oslo_vmware.api [None req-e02d4b37-ae77-44da-acd2-da4ff1500a31 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Task: {'id': task-1313951, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.232983] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0aa06aad-3d54-4fa6-8a41-6527ef18a645 tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Lock "003191d3-fe38-4bde-8d7c-75b4aad413b4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.232s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.300832] env[62460]: DEBUG oslo_vmware.api [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313949, 'name': PowerOnVM_Task, 'duration_secs': 0.57008} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.302071] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 963.302167] env[62460]: INFO nova.compute.manager [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Took 7.96 seconds to spawn the instance on the hypervisor. [ 963.302361] env[62460]: DEBUG nova.compute.manager [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 963.303161] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e58ab052-af52-445f-ba6c-379487031c11 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.335430] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313950, 'name': CreateVM_Task, 'duration_secs': 0.311252} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.335566] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 963.336750] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.336750] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.336859] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 963.339323] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-43719d79-2e74-46de-bf1f-da4f0192857d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.347695] env[62460]: DEBUG oslo_vmware.api [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for the task: (returnval){ [ 963.347695] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]529bbb31-93f7-c996-84d0-6face6c6f985" [ 963.347695] env[62460]: _type = "Task" [ 963.347695] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.352790] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acf6d7ce-ad32-416e-89af-5acc6e13fe8e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.361779] env[62460]: DEBUG oslo_vmware.api [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]529bbb31-93f7-c996-84d0-6face6c6f985, 'name': SearchDatastore_Task, 'duration_secs': 0.011352} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.364307] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.364623] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 963.365780] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.366087] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.366286] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 963.366822] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8cd9af37-a942-4a92-b8fe-a7bc3f0cc26d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.370238] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82176d5d-abc7-4633-a352-68a5af39bf4f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.425258] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-482310c7-41ec-4c3d-9ab8-0706e6c36e28 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.429437] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 963.429697] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 963.434028] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-40bae6f0-2b74-44c6-ac72-b281bf7a6b45 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.439834] env[62460]: DEBUG oslo_vmware.api [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for the task: (returnval){ [ 963.439834] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]522984d8-1fe7-2134-71c8-434218f43b1b" [ 963.439834] env[62460]: _type = "Task" [ 963.439834] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.441272] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33e8bdcd-db71-4462-b535-014d4afb9ec3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.457428] env[62460]: DEBUG oslo_vmware.api [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]522984d8-1fe7-2134-71c8-434218f43b1b, 'name': SearchDatastore_Task, 'duration_secs': 0.009429} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.465904] env[62460]: DEBUG nova.compute.provider_tree [None req-0fa3495d-783c-4299-a9d5-dee125e75d0d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 963.467160] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e74a965b-a0bb-4b73-845f-25cb428d046d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.472730] env[62460]: DEBUG oslo_vmware.api [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for the task: (returnval){ [ 963.472730] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52e8ac42-1f26-8395-1c21-465c0d5ffdc3" [ 963.472730] env[62460]: _type = "Task" [ 963.472730] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.484111] env[62460]: DEBUG oslo_vmware.api [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52e8ac42-1f26-8395-1c21-465c0d5ffdc3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.581184] env[62460]: DEBUG nova.network.neutron [req-b855487b-cf08-43f5-bbe3-d7feb8b295ba req-5124bc5c-fe94-467c-b6b8-80a216d82abf service nova] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Updated VIF entry in instance network info cache for port 352f5563-2248-439c-a843-2001a2d5d230. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 963.581676] env[62460]: DEBUG nova.network.neutron [req-b855487b-cf08-43f5-bbe3-d7feb8b295ba req-5124bc5c-fe94-467c-b6b8-80a216d82abf service nova] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Updating instance_info_cache with network_info: [{"id": "352f5563-2248-439c-a843-2001a2d5d230", "address": "fa:16:3e:fd:52:b2", "network": {"id": "c8b3d824-c8f0-4f7b-a91f-1987e04a5f09", "bridge": "br-int", "label": "tempest-ServersTestJSON-1769082812-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10ac54db1ef54e249a077bbb3dde4242", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "255460d5-71d4-4bfd-87f1-acc10085db7f", "external-id": "nsx-vlan-transportzone-152", "segmentation_id": 152, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap352f5563-22", "ovs_interfaceid": "352f5563-2248-439c-a843-2001a2d5d230", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.681798] env[62460]: DEBUG oslo_vmware.api [None req-e02d4b37-ae77-44da-acd2-da4ff1500a31 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Task: {'id': task-1313951, 'name': ReconfigVM_Task, 'duration_secs': 0.150652} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.682088] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-e02d4b37-ae77-44da-acd2-da4ff1500a31 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Reconfigured VM instance to set the machine id {{(pid=62460) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1799}} [ 963.745303] env[62460]: DEBUG nova.network.neutron [None req-1de5df41-5d11-43c8-9708-d4900076226b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Updating instance_info_cache with network_info: [{"id": "546b0347-51db-4a4a-99e9-4f3db96a36bd", "address": "fa:16:3e:54:9b:7b", "network": {"id": "d3136e32-ad55-4b73-835c-8fa5f0480767", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1511414968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53fa06008e9f43488362895e7a143700", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "054fcd1e-638e-425a-a1de-78cb188ae026", "external-id": "nsx-vlan-transportzone-658", "segmentation_id": 658, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap546b0347-51", "ovs_interfaceid": "546b0347-51db-4a4a-99e9-4f3db96a36bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "50d6a0c4-3681-4c1a-b42f-2987ee7a3bfb", "address": "fa:16:3e:c6:6b:1c", "network": {"id": "d3136e32-ad55-4b73-835c-8fa5f0480767", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1511414968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53fa06008e9f43488362895e7a143700", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "054fcd1e-638e-425a-a1de-78cb188ae026", "external-id": "nsx-vlan-transportzone-658", "segmentation_id": 658, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap50d6a0c4-36", "ovs_interfaceid": "50d6a0c4-3681-4c1a-b42f-2987ee7a3bfb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.809161] env[62460]: DEBUG nova.network.neutron [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Successfully updated port: 786ef177-843f-44d0-8920-2f4332e29154 {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 963.828736] env[62460]: INFO nova.compute.manager [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Took 15.55 seconds to build instance. [ 963.904213] env[62460]: DEBUG nova.compute.manager [req-f354731f-ed2c-424d-b632-b87ff50f6f3c req-0c6dec7c-6e0d-420c-8e3f-6d1017c8a930 service nova] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Received event network-changed-7bb00d79-430f-44e2-9e92-57109c05d0bb {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 963.904432] env[62460]: DEBUG nova.compute.manager [req-f354731f-ed2c-424d-b632-b87ff50f6f3c req-0c6dec7c-6e0d-420c-8e3f-6d1017c8a930 service nova] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Refreshing instance network info cache due to event network-changed-7bb00d79-430f-44e2-9e92-57109c05d0bb. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 963.904657] env[62460]: DEBUG oslo_concurrency.lockutils [req-f354731f-ed2c-424d-b632-b87ff50f6f3c req-0c6dec7c-6e0d-420c-8e3f-6d1017c8a930 service nova] Acquiring lock "refresh_cache-bf28fc98-d9a9-4678-a107-ace2b6503353" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.904805] env[62460]: DEBUG oslo_concurrency.lockutils [req-f354731f-ed2c-424d-b632-b87ff50f6f3c req-0c6dec7c-6e0d-420c-8e3f-6d1017c8a930 service nova] Acquired lock "refresh_cache-bf28fc98-d9a9-4678-a107-ace2b6503353" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.904972] env[62460]: DEBUG nova.network.neutron [req-f354731f-ed2c-424d-b632-b87ff50f6f3c req-0c6dec7c-6e0d-420c-8e3f-6d1017c8a930 service nova] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Refreshing network info cache for port 7bb00d79-430f-44e2-9e92-57109c05d0bb {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 963.970474] env[62460]: DEBUG nova.scheduler.client.report [None req-0fa3495d-783c-4299-a9d5-dee125e75d0d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 963.985612] env[62460]: DEBUG oslo_vmware.api [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52e8ac42-1f26-8395-1c21-465c0d5ffdc3, 'name': SearchDatastore_Task, 'duration_secs': 0.029867} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.986332] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.986595] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] f0f66557-04ef-4a96-8a56-e640bfda3b36/f0f66557-04ef-4a96-8a56-e640bfda3b36.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 963.986877] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7ba58c58-eec5-4892-b28e-2ef7b89dd792 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.992820] env[62460]: DEBUG oslo_vmware.api [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for the task: (returnval){ [ 963.992820] env[62460]: value = "task-1313952" [ 963.992820] env[62460]: _type = "Task" [ 963.992820] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.000278] env[62460]: DEBUG oslo_vmware.api [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313952, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.078231] env[62460]: DEBUG oslo_concurrency.lockutils [None req-189c95d3-7540-4455-870e-38113115b367 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Acquiring lock "bf28fc98-d9a9-4678-a107-ace2b6503353" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.078481] env[62460]: DEBUG oslo_concurrency.lockutils [None req-189c95d3-7540-4455-870e-38113115b367 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Lock "bf28fc98-d9a9-4678-a107-ace2b6503353" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.078697] env[62460]: DEBUG oslo_concurrency.lockutils [None req-189c95d3-7540-4455-870e-38113115b367 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Acquiring lock "bf28fc98-d9a9-4678-a107-ace2b6503353-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.078892] env[62460]: DEBUG oslo_concurrency.lockutils [None req-189c95d3-7540-4455-870e-38113115b367 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Lock "bf28fc98-d9a9-4678-a107-ace2b6503353-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.079223] env[62460]: DEBUG oslo_concurrency.lockutils [None req-189c95d3-7540-4455-870e-38113115b367 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Lock "bf28fc98-d9a9-4678-a107-ace2b6503353-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.081391] env[62460]: INFO nova.compute.manager [None req-189c95d3-7540-4455-870e-38113115b367 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Terminating instance [ 964.083143] env[62460]: DEBUG nova.compute.manager [None req-189c95d3-7540-4455-870e-38113115b367 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 964.083345] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-189c95d3-7540-4455-870e-38113115b367 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 964.083815] env[62460]: DEBUG oslo_concurrency.lockutils [req-b855487b-cf08-43f5-bbe3-d7feb8b295ba req-5124bc5c-fe94-467c-b6b8-80a216d82abf service nova] Releasing lock "refresh_cache-f0f66557-04ef-4a96-8a56-e640bfda3b36" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.084682] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf4f9cbd-1da8-4f13-adb4-0a12685df8a1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.092133] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-189c95d3-7540-4455-870e-38113115b367 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 964.092370] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8c44d85c-e0a0-4ed6-8cd3-ea8df9cd14e2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.098496] env[62460]: DEBUG oslo_vmware.api [None req-189c95d3-7540-4455-870e-38113115b367 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Waiting for the task: (returnval){ [ 964.098496] env[62460]: value = "task-1313953" [ 964.098496] env[62460]: _type = "Task" [ 964.098496] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.107359] env[62460]: DEBUG oslo_vmware.api [None req-189c95d3-7540-4455-870e-38113115b367 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Task: {'id': task-1313953, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.169677] env[62460]: DEBUG nova.compute.manager [req-600ab554-77ef-409a-ace1-d53b8461c434 req-fd3be2bf-81fe-4a14-99ee-bbefb0951690 service nova] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Received event network-changed-50d6a0c4-3681-4c1a-b42f-2987ee7a3bfb {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 964.169955] env[62460]: DEBUG nova.compute.manager [req-600ab554-77ef-409a-ace1-d53b8461c434 req-fd3be2bf-81fe-4a14-99ee-bbefb0951690 service nova] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Refreshing instance network info cache due to event network-changed-50d6a0c4-3681-4c1a-b42f-2987ee7a3bfb. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 964.170712] env[62460]: DEBUG oslo_concurrency.lockutils [req-600ab554-77ef-409a-ace1-d53b8461c434 req-fd3be2bf-81fe-4a14-99ee-bbefb0951690 service nova] Acquiring lock "refresh_cache-abd4501b-3ad8-4103-921d-5b80fc2f313f" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 964.248411] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1de5df41-5d11-43c8-9708-d4900076226b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Releasing lock "refresh_cache-abd4501b-3ad8-4103-921d-5b80fc2f313f" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.249218] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1de5df41-5d11-43c8-9708-d4900076226b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "abd4501b-3ad8-4103-921d-5b80fc2f313f" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 964.249455] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1de5df41-5d11-43c8-9708-d4900076226b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquired lock "abd4501b-3ad8-4103-921d-5b80fc2f313f" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.250116] env[62460]: DEBUG oslo_concurrency.lockutils [req-600ab554-77ef-409a-ace1-d53b8461c434 req-fd3be2bf-81fe-4a14-99ee-bbefb0951690 service nova] Acquired lock "refresh_cache-abd4501b-3ad8-4103-921d-5b80fc2f313f" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.250359] env[62460]: DEBUG nova.network.neutron [req-600ab554-77ef-409a-ace1-d53b8461c434 req-fd3be2bf-81fe-4a14-99ee-bbefb0951690 service nova] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Refreshing network info cache for port 50d6a0c4-3681-4c1a-b42f-2987ee7a3bfb {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 964.252190] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35284ef4-8db4-49b4-a87a-532922d8a9a3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.274058] env[62460]: DEBUG nova.virt.hardware [None req-1de5df41-5d11-43c8-9708-d4900076226b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 964.274369] env[62460]: DEBUG nova.virt.hardware [None req-1de5df41-5d11-43c8-9708-d4900076226b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 964.274550] env[62460]: DEBUG nova.virt.hardware [None req-1de5df41-5d11-43c8-9708-d4900076226b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 964.274724] env[62460]: DEBUG nova.virt.hardware [None req-1de5df41-5d11-43c8-9708-d4900076226b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 964.274899] env[62460]: DEBUG nova.virt.hardware [None req-1de5df41-5d11-43c8-9708-d4900076226b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 964.275370] env[62460]: DEBUG nova.virt.hardware [None req-1de5df41-5d11-43c8-9708-d4900076226b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 964.275678] env[62460]: DEBUG nova.virt.hardware [None req-1de5df41-5d11-43c8-9708-d4900076226b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 964.275883] env[62460]: DEBUG nova.virt.hardware [None req-1de5df41-5d11-43c8-9708-d4900076226b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 964.276113] env[62460]: DEBUG nova.virt.hardware [None req-1de5df41-5d11-43c8-9708-d4900076226b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 964.276300] env[62460]: DEBUG nova.virt.hardware [None req-1de5df41-5d11-43c8-9708-d4900076226b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 964.276729] env[62460]: DEBUG nova.virt.hardware [None req-1de5df41-5d11-43c8-9708-d4900076226b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 964.283817] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-1de5df41-5d11-43c8-9708-d4900076226b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Reconfiguring VM to attach interface {{(pid=62460) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 964.284980] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fd6fd046-25f0-42d3-b164-28c5a801f352 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.303624] env[62460]: DEBUG oslo_vmware.api [None req-1de5df41-5d11-43c8-9708-d4900076226b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Waiting for the task: (returnval){ [ 964.303624] env[62460]: value = "task-1313954" [ 964.303624] env[62460]: _type = "Task" [ 964.303624] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.312585] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "refresh_cache-0d39a304-6a49-4d87-bfa2-4ba3b383578a" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 964.312723] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquired lock "refresh_cache-0d39a304-6a49-4d87-bfa2-4ba3b383578a" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.312869] env[62460]: DEBUG nova.network.neutron [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 964.313979] env[62460]: DEBUG oslo_vmware.api [None req-1de5df41-5d11-43c8-9708-d4900076226b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313954, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.331036] env[62460]: DEBUG oslo_concurrency.lockutils [None req-39dd44a6-cda1-4a76-8bb2-9aa820b19a5a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "290500b5-d5b4-43b6-a843-20dacf561f4b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.063s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.475710] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0fa3495d-783c-4299-a9d5-dee125e75d0d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.773s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.501324] env[62460]: INFO nova.scheduler.client.report [None req-0fa3495d-783c-4299-a9d5-dee125e75d0d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Deleted allocations for instance db331001-1744-4c42-a40c-87f3e60da4b8 [ 964.506233] env[62460]: DEBUG oslo_vmware.api [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313952, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.608729] env[62460]: DEBUG oslo_vmware.api [None req-189c95d3-7540-4455-870e-38113115b367 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Task: {'id': task-1313953, 'name': PowerOffVM_Task, 'duration_secs': 0.186824} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.608987] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-189c95d3-7540-4455-870e-38113115b367 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 964.609316] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-189c95d3-7540-4455-870e-38113115b367 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 964.609605] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cc57d068-5312-4813-9457-7e7ba57be9ca {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.674129] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-189c95d3-7540-4455-870e-38113115b367 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 964.674398] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-189c95d3-7540-4455-870e-38113115b367 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Deleting contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 964.674593] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-189c95d3-7540-4455-870e-38113115b367 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Deleting the datastore file [datastore2] bf28fc98-d9a9-4678-a107-ace2b6503353 {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 964.674856] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d388bd51-b66a-4ae2-9b38-db552679a908 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.681223] env[62460]: DEBUG oslo_vmware.api [None req-189c95d3-7540-4455-870e-38113115b367 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Waiting for the task: (returnval){ [ 964.681223] env[62460]: value = "task-1313956" [ 964.681223] env[62460]: _type = "Task" [ 964.681223] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.690213] env[62460]: DEBUG oslo_vmware.api [None req-189c95d3-7540-4455-870e-38113115b367 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Task: {'id': task-1313956, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.706624] env[62460]: DEBUG nova.network.neutron [req-f354731f-ed2c-424d-b632-b87ff50f6f3c req-0c6dec7c-6e0d-420c-8e3f-6d1017c8a930 service nova] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Updated VIF entry in instance network info cache for port 7bb00d79-430f-44e2-9e92-57109c05d0bb. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 964.707013] env[62460]: DEBUG nova.network.neutron [req-f354731f-ed2c-424d-b632-b87ff50f6f3c req-0c6dec7c-6e0d-420c-8e3f-6d1017c8a930 service nova] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Updating instance_info_cache with network_info: [{"id": "7bb00d79-430f-44e2-9e92-57109c05d0bb", "address": "fa:16:3e:94:48:0c", "network": {"id": "e2ad0587-12a2-40ce-b2d0-c682042335fb", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1549110976-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.202", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1845e17fb7e64cf594c4e81c67ad8243", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c405e9f-a6c8-4308-acac-071654efe18e", "external-id": "nsx-vlan-transportzone-851", "segmentation_id": 851, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7bb00d79-43", "ovs_interfaceid": "7bb00d79-430f-44e2-9e92-57109c05d0bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 964.813657] env[62460]: DEBUG oslo_vmware.api [None req-1de5df41-5d11-43c8-9708-d4900076226b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313954, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.879153] env[62460]: DEBUG nova.network.neutron [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 965.003868] env[62460]: DEBUG oslo_vmware.api [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313952, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.698673} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.004225] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] f0f66557-04ef-4a96-8a56-e640bfda3b36/f0f66557-04ef-4a96-8a56-e640bfda3b36.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 965.004467] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 965.004728] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0ebcba11-ddbc-4063-9e8a-023c54177066 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.011953] env[62460]: DEBUG oslo_vmware.api [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for the task: (returnval){ [ 965.011953] env[62460]: value = "task-1313957" [ 965.011953] env[62460]: _type = "Task" [ 965.011953] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.012405] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0fa3495d-783c-4299-a9d5-dee125e75d0d tempest-ListServersNegativeTestJSON-1241582871 tempest-ListServersNegativeTestJSON-1241582871-project-member] Lock "db331001-1744-4c42-a40c-87f3e60da4b8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.945s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.021686] env[62460]: DEBUG oslo_vmware.api [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313957, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.195630] env[62460]: DEBUG oslo_vmware.api [None req-189c95d3-7540-4455-870e-38113115b367 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Task: {'id': task-1313956, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.40975} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.195984] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-189c95d3-7540-4455-870e-38113115b367 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 965.196390] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-189c95d3-7540-4455-870e-38113115b367 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Deleted contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 965.196572] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-189c95d3-7540-4455-870e-38113115b367 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 965.196766] env[62460]: INFO nova.compute.manager [None req-189c95d3-7540-4455-870e-38113115b367 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Took 1.11 seconds to destroy the instance on the hypervisor. [ 965.197038] env[62460]: DEBUG oslo.service.loopingcall [None req-189c95d3-7540-4455-870e-38113115b367 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 965.197274] env[62460]: DEBUG nova.compute.manager [-] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 965.197368] env[62460]: DEBUG nova.network.neutron [-] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 965.209514] env[62460]: DEBUG oslo_concurrency.lockutils [req-f354731f-ed2c-424d-b632-b87ff50f6f3c req-0c6dec7c-6e0d-420c-8e3f-6d1017c8a930 service nova] Releasing lock "refresh_cache-bf28fc98-d9a9-4678-a107-ace2b6503353" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.209802] env[62460]: DEBUG nova.compute.manager [req-f354731f-ed2c-424d-b632-b87ff50f6f3c req-0c6dec7c-6e0d-420c-8e3f-6d1017c8a930 service nova] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Received event network-vif-plugged-786ef177-843f-44d0-8920-2f4332e29154 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 965.209998] env[62460]: DEBUG oslo_concurrency.lockutils [req-f354731f-ed2c-424d-b632-b87ff50f6f3c req-0c6dec7c-6e0d-420c-8e3f-6d1017c8a930 service nova] Acquiring lock "0d39a304-6a49-4d87-bfa2-4ba3b383578a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.210262] env[62460]: DEBUG oslo_concurrency.lockutils [req-f354731f-ed2c-424d-b632-b87ff50f6f3c req-0c6dec7c-6e0d-420c-8e3f-6d1017c8a930 service nova] Lock "0d39a304-6a49-4d87-bfa2-4ba3b383578a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.210451] env[62460]: DEBUG oslo_concurrency.lockutils [req-f354731f-ed2c-424d-b632-b87ff50f6f3c req-0c6dec7c-6e0d-420c-8e3f-6d1017c8a930 service nova] Lock "0d39a304-6a49-4d87-bfa2-4ba3b383578a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.210630] env[62460]: DEBUG nova.compute.manager [req-f354731f-ed2c-424d-b632-b87ff50f6f3c req-0c6dec7c-6e0d-420c-8e3f-6d1017c8a930 service nova] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] No waiting events found dispatching network-vif-plugged-786ef177-843f-44d0-8920-2f4332e29154 {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 965.210808] env[62460]: WARNING nova.compute.manager [req-f354731f-ed2c-424d-b632-b87ff50f6f3c req-0c6dec7c-6e0d-420c-8e3f-6d1017c8a930 service nova] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Received unexpected event network-vif-plugged-786ef177-843f-44d0-8920-2f4332e29154 for instance with vm_state building and task_state spawning. [ 965.211122] env[62460]: DEBUG nova.compute.manager [req-f354731f-ed2c-424d-b632-b87ff50f6f3c req-0c6dec7c-6e0d-420c-8e3f-6d1017c8a930 service nova] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Received event network-changed-786ef177-843f-44d0-8920-2f4332e29154 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 965.211240] env[62460]: DEBUG nova.compute.manager [req-f354731f-ed2c-424d-b632-b87ff50f6f3c req-0c6dec7c-6e0d-420c-8e3f-6d1017c8a930 service nova] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Refreshing instance network info cache due to event network-changed-786ef177-843f-44d0-8920-2f4332e29154. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 965.211314] env[62460]: DEBUG oslo_concurrency.lockutils [req-f354731f-ed2c-424d-b632-b87ff50f6f3c req-0c6dec7c-6e0d-420c-8e3f-6d1017c8a930 service nova] Acquiring lock "refresh_cache-0d39a304-6a49-4d87-bfa2-4ba3b383578a" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 965.227120] env[62460]: DEBUG nova.network.neutron [req-600ab554-77ef-409a-ace1-d53b8461c434 req-fd3be2bf-81fe-4a14-99ee-bbefb0951690 service nova] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Updated VIF entry in instance network info cache for port 50d6a0c4-3681-4c1a-b42f-2987ee7a3bfb. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 965.227575] env[62460]: DEBUG nova.network.neutron [req-600ab554-77ef-409a-ace1-d53b8461c434 req-fd3be2bf-81fe-4a14-99ee-bbefb0951690 service nova] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Updating instance_info_cache with network_info: [{"id": "546b0347-51db-4a4a-99e9-4f3db96a36bd", "address": "fa:16:3e:54:9b:7b", "network": {"id": "d3136e32-ad55-4b73-835c-8fa5f0480767", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1511414968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53fa06008e9f43488362895e7a143700", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "054fcd1e-638e-425a-a1de-78cb188ae026", "external-id": "nsx-vlan-transportzone-658", "segmentation_id": 658, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap546b0347-51", "ovs_interfaceid": "546b0347-51db-4a4a-99e9-4f3db96a36bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "50d6a0c4-3681-4c1a-b42f-2987ee7a3bfb", "address": "fa:16:3e:c6:6b:1c", "network": {"id": "d3136e32-ad55-4b73-835c-8fa5f0480767", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1511414968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53fa06008e9f43488362895e7a143700", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "054fcd1e-638e-425a-a1de-78cb188ae026", "external-id": "nsx-vlan-transportzone-658", "segmentation_id": 658, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap50d6a0c4-36", "ovs_interfaceid": "50d6a0c4-3681-4c1a-b42f-2987ee7a3bfb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 965.230960] env[62460]: DEBUG nova.network.neutron [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Updating instance_info_cache with network_info: [{"id": "786ef177-843f-44d0-8920-2f4332e29154", "address": "fa:16:3e:09:59:ab", "network": {"id": "3c49720c-94c9-4e2f-bbb7-4266cc97faee", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1650281790-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfa5a65b0f614e769de5b3aa77bf869e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap786ef177-84", "ovs_interfaceid": "786ef177-843f-44d0-8920-2f4332e29154", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 965.314407] env[62460]: DEBUG oslo_vmware.api [None req-1de5df41-5d11-43c8-9708-d4900076226b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313954, 'name': ReconfigVM_Task, 'duration_secs': 0.978473} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.314985] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1de5df41-5d11-43c8-9708-d4900076226b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Releasing lock "abd4501b-3ad8-4103-921d-5b80fc2f313f" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.315256] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-1de5df41-5d11-43c8-9708-d4900076226b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Reconfigured VM to attach interface {{(pid=62460) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 965.523456] env[62460]: DEBUG oslo_vmware.api [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313957, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.097498} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.523737] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 965.524521] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-643cf424-53b4-40cc-84ef-7f67d2599163 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.546714] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Reconfiguring VM instance instance-0000005f to attach disk [datastore1] f0f66557-04ef-4a96-8a56-e640bfda3b36/f0f66557-04ef-4a96-8a56-e640bfda3b36.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 965.546999] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9b2c5497-5874-4a4a-9e27-95031ca5335c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.565956] env[62460]: DEBUG oslo_vmware.api [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for the task: (returnval){ [ 965.565956] env[62460]: value = "task-1313958" [ 965.565956] env[62460]: _type = "Task" [ 965.565956] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.573428] env[62460]: DEBUG oslo_vmware.api [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313958, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.730492] env[62460]: DEBUG oslo_concurrency.lockutils [req-600ab554-77ef-409a-ace1-d53b8461c434 req-fd3be2bf-81fe-4a14-99ee-bbefb0951690 service nova] Releasing lock "refresh_cache-abd4501b-3ad8-4103-921d-5b80fc2f313f" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.733324] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Releasing lock "refresh_cache-0d39a304-6a49-4d87-bfa2-4ba3b383578a" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.733706] env[62460]: DEBUG nova.compute.manager [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Instance network_info: |[{"id": "786ef177-843f-44d0-8920-2f4332e29154", "address": "fa:16:3e:09:59:ab", "network": {"id": "3c49720c-94c9-4e2f-bbb7-4266cc97faee", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1650281790-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfa5a65b0f614e769de5b3aa77bf869e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap786ef177-84", "ovs_interfaceid": "786ef177-843f-44d0-8920-2f4332e29154", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 965.734026] env[62460]: DEBUG oslo_concurrency.lockutils [req-f354731f-ed2c-424d-b632-b87ff50f6f3c req-0c6dec7c-6e0d-420c-8e3f-6d1017c8a930 service nova] Acquired lock "refresh_cache-0d39a304-6a49-4d87-bfa2-4ba3b383578a" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.734221] env[62460]: DEBUG nova.network.neutron [req-f354731f-ed2c-424d-b632-b87ff50f6f3c req-0c6dec7c-6e0d-420c-8e3f-6d1017c8a930 service nova] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Refreshing network info cache for port 786ef177-843f-44d0-8920-2f4332e29154 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 965.739024] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:09:59:ab', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '04ccbc7a-cf8d-4ea2-8411-291a1e27df7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '786ef177-843f-44d0-8920-2f4332e29154', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 965.744594] env[62460]: DEBUG oslo.service.loopingcall [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 965.747669] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 965.748268] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-030877b0-38dd-4a0b-bb1e-f0511babf3f8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.768378] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 965.768378] env[62460]: value = "task-1313959" [ 965.768378] env[62460]: _type = "Task" [ 965.768378] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.776281] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313959, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.819676] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1de5df41-5d11-43c8-9708-d4900076226b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "interface-abd4501b-3ad8-4103-921d-5b80fc2f313f-50d6a0c4-3681-4c1a-b42f-2987ee7a3bfb" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.366s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.977696] env[62460]: DEBUG nova.network.neutron [req-f354731f-ed2c-424d-b632-b87ff50f6f3c req-0c6dec7c-6e0d-420c-8e3f-6d1017c8a930 service nova] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Updated VIF entry in instance network info cache for port 786ef177-843f-44d0-8920-2f4332e29154. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 965.978144] env[62460]: DEBUG nova.network.neutron [req-f354731f-ed2c-424d-b632-b87ff50f6f3c req-0c6dec7c-6e0d-420c-8e3f-6d1017c8a930 service nova] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Updating instance_info_cache with network_info: [{"id": "786ef177-843f-44d0-8920-2f4332e29154", "address": "fa:16:3e:09:59:ab", "network": {"id": "3c49720c-94c9-4e2f-bbb7-4266cc97faee", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1650281790-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfa5a65b0f614e769de5b3aa77bf869e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap786ef177-84", "ovs_interfaceid": "786ef177-843f-44d0-8920-2f4332e29154", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.077487] env[62460]: DEBUG oslo_vmware.api [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313958, 'name': ReconfigVM_Task, 'duration_secs': 0.327529} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.077782] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Reconfigured VM instance instance-0000005f to attach disk [datastore1] f0f66557-04ef-4a96-8a56-e640bfda3b36/f0f66557-04ef-4a96-8a56-e640bfda3b36.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 966.079758] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-28ad0bf1-36db-494e-a03b-05e0fcac9c4a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.088469] env[62460]: DEBUG oslo_vmware.api [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for the task: (returnval){ [ 966.088469] env[62460]: value = "task-1313960" [ 966.088469] env[62460]: _type = "Task" [ 966.088469] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.100451] env[62460]: DEBUG oslo_vmware.api [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313960, 'name': Rename_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.281763] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313959, 'name': CreateVM_Task, 'duration_secs': 0.34907} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.282081] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 966.282741] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.282922] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.283685] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 966.283685] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-44d0fdf7-a9fb-4174-8186-89b35cf4ddb0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.288779] env[62460]: DEBUG oslo_vmware.api [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 966.288779] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52262976-5f13-1428-b512-df2db398a34b" [ 966.288779] env[62460]: _type = "Task" [ 966.288779] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.298058] env[62460]: DEBUG oslo_vmware.api [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52262976-5f13-1428-b512-df2db398a34b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.311197] env[62460]: DEBUG nova.compute.manager [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Stashing vm_state: active {{(pid=62460) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 966.344850] env[62460]: DEBUG nova.compute.manager [req-441e3f43-aee2-4441-b4b2-4dc60f24bdce req-e6a360dc-4d73-40fd-9dc2-b66b4c88b4ef service nova] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Received event network-vif-deleted-7bb00d79-430f-44e2-9e92-57109c05d0bb {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 966.345724] env[62460]: INFO nova.compute.manager [req-441e3f43-aee2-4441-b4b2-4dc60f24bdce req-e6a360dc-4d73-40fd-9dc2-b66b4c88b4ef service nova] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Neutron deleted interface 7bb00d79-430f-44e2-9e92-57109c05d0bb; detaching it from the instance and deleting it from the info cache [ 966.345724] env[62460]: DEBUG nova.network.neutron [req-441e3f43-aee2-4441-b4b2-4dc60f24bdce req-e6a360dc-4d73-40fd-9dc2-b66b4c88b4ef service nova] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.436435] env[62460]: DEBUG nova.network.neutron [-] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.480551] env[62460]: DEBUG oslo_concurrency.lockutils [req-f354731f-ed2c-424d-b632-b87ff50f6f3c req-0c6dec7c-6e0d-420c-8e3f-6d1017c8a930 service nova] Releasing lock "refresh_cache-0d39a304-6a49-4d87-bfa2-4ba3b383578a" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.600269] env[62460]: DEBUG oslo_vmware.api [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313960, 'name': Rename_Task} progress is 99%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.799892] env[62460]: DEBUG oslo_vmware.api [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52262976-5f13-1428-b512-df2db398a34b, 'name': SearchDatastore_Task, 'duration_secs': 0.023675} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.800302] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.800601] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 966.800876] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.801042] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.801232] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 966.801601] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-45526d61-bd7d-4e01-8e00-c7b8a2acc4d5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.811734] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 966.811882] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 966.812617] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-273f3649-2879-40bd-8a3f-c00ce1567fa1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.821162] env[62460]: DEBUG oslo_vmware.api [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 966.821162] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52fd99d6-a2c1-a8ed-caac-3a5685157b3a" [ 966.821162] env[62460]: _type = "Task" [ 966.821162] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.835446] env[62460]: DEBUG oslo_vmware.api [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52fd99d6-a2c1-a8ed-caac-3a5685157b3a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.840106] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.840419] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.850934] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fcb8e89d-1f92-4ed4-9441-d81f59f8a58a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.862177] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0842feca-7749-4f3e-a664-6d053582c768 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.891186] env[62460]: DEBUG nova.compute.manager [req-441e3f43-aee2-4441-b4b2-4dc60f24bdce req-e6a360dc-4d73-40fd-9dc2-b66b4c88b4ef service nova] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Detach interface failed, port_id=7bb00d79-430f-44e2-9e92-57109c05d0bb, reason: Instance bf28fc98-d9a9-4678-a107-ace2b6503353 could not be found. {{(pid=62460) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 966.940983] env[62460]: INFO nova.compute.manager [-] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Took 1.74 seconds to deallocate network for instance. [ 967.099015] env[62460]: DEBUG oslo_vmware.api [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313960, 'name': Rename_Task} progress is 99%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.156478] env[62460]: DEBUG oslo_concurrency.lockutils [None req-daf435a1-8ec6-40b0-b5b0-b170e8f02000 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "interface-abd4501b-3ad8-4103-921d-5b80fc2f313f-50d6a0c4-3681-4c1a-b42f-2987ee7a3bfb" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.156790] env[62460]: DEBUG oslo_concurrency.lockutils [None req-daf435a1-8ec6-40b0-b5b0-b170e8f02000 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "interface-abd4501b-3ad8-4103-921d-5b80fc2f313f-50d6a0c4-3681-4c1a-b42f-2987ee7a3bfb" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.332115] env[62460]: DEBUG oslo_vmware.api [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52fd99d6-a2c1-a8ed-caac-3a5685157b3a, 'name': SearchDatastore_Task, 'duration_secs': 0.020602} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.333051] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-56297ac1-1b5f-4876-8ffd-f9913c065e76 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.338640] env[62460]: DEBUG oslo_vmware.api [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 967.338640] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]5255ff4d-abfb-02bd-09c9-ff5b3bf15fc0" [ 967.338640] env[62460]: _type = "Task" [ 967.338640] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.346367] env[62460]: INFO nova.compute.claims [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 967.356060] env[62460]: DEBUG oslo_vmware.api [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5255ff4d-abfb-02bd-09c9-ff5b3bf15fc0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.449687] env[62460]: DEBUG oslo_concurrency.lockutils [None req-189c95d3-7540-4455-870e-38113115b367 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.600585] env[62460]: DEBUG oslo_vmware.api [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313960, 'name': Rename_Task, 'duration_secs': 1.147793} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.600875] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 967.601143] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3d627d66-fddb-482f-9b70-c3484baef1dc {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.607011] env[62460]: DEBUG oslo_vmware.api [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for the task: (returnval){ [ 967.607011] env[62460]: value = "task-1313961" [ 967.607011] env[62460]: _type = "Task" [ 967.607011] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.614077] env[62460]: DEBUG oslo_vmware.api [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313961, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.659497] env[62460]: DEBUG oslo_concurrency.lockutils [None req-daf435a1-8ec6-40b0-b5b0-b170e8f02000 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "abd4501b-3ad8-4103-921d-5b80fc2f313f" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.659703] env[62460]: DEBUG oslo_concurrency.lockutils [None req-daf435a1-8ec6-40b0-b5b0-b170e8f02000 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquired lock "abd4501b-3ad8-4103-921d-5b80fc2f313f" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.660623] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f43d5b15-8c67-41c7-bb84-24da213ca51a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.677516] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-145c9e5e-0632-4fe3-9908-0bbef9cb2758 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.709350] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-daf435a1-8ec6-40b0-b5b0-b170e8f02000 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Reconfiguring VM to detach interface {{(pid=62460) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 967.712019] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b116fae1-cfdf-447a-92f5-cb76fc74bea8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.734146] env[62460]: DEBUG oslo_vmware.api [None req-daf435a1-8ec6-40b0-b5b0-b170e8f02000 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Waiting for the task: (returnval){ [ 967.734146] env[62460]: value = "task-1313962" [ 967.734146] env[62460]: _type = "Task" [ 967.734146] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.746153] env[62460]: DEBUG oslo_vmware.api [None req-daf435a1-8ec6-40b0-b5b0-b170e8f02000 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313962, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.849737] env[62460]: DEBUG oslo_vmware.api [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5255ff4d-abfb-02bd-09c9-ff5b3bf15fc0, 'name': SearchDatastore_Task, 'duration_secs': 0.011413} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.849737] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.850021] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 0d39a304-6a49-4d87-bfa2-4ba3b383578a/0d39a304-6a49-4d87-bfa2-4ba3b383578a.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 967.850327] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6299a6ae-47ed-412c-99fd-b070c3c46ce0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.858604] env[62460]: INFO nova.compute.resource_tracker [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Updating resource usage from migration 6c432a03-9e05-49cc-bf18-7f0bc8bde6eb [ 967.860575] env[62460]: DEBUG oslo_vmware.api [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 967.860575] env[62460]: value = "task-1313963" [ 967.860575] env[62460]: _type = "Task" [ 967.860575] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.871608] env[62460]: DEBUG oslo_vmware.api [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313963, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.002153] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09dc0280-1c66-449b-8b3b-a2ec6d958a9b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.009436] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b1fdde3-8c2e-4a61-9135-61a24274090c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.039669] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be85a2e4-dc6e-4d28-b64f-b035593f8877 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.047469] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34ea58ca-9cbc-4287-ba93-16af3e016bb7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.061370] env[62460]: DEBUG nova.compute.provider_tree [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 968.117631] env[62460]: DEBUG oslo_vmware.api [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313961, 'name': PowerOnVM_Task} progress is 94%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.244436] env[62460]: DEBUG oslo_vmware.api [None req-daf435a1-8ec6-40b0-b5b0-b170e8f02000 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313962, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.371582] env[62460]: DEBUG oslo_vmware.api [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313963, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.588371] env[62460]: ERROR nova.scheduler.client.report [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [req-83559fd1-890c-4b8c-a216-42069ca82d6a] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 76e8f00d-65cc-4766-bb9d-5006fb51c728. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-83559fd1-890c-4b8c-a216-42069ca82d6a"}]} [ 968.616635] env[62460]: DEBUG nova.scheduler.client.report [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Refreshing inventories for resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 968.624600] env[62460]: DEBUG oslo_vmware.api [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313961, 'name': PowerOnVM_Task, 'duration_secs': 0.733851} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.624964] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 968.625255] env[62460]: INFO nova.compute.manager [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Took 8.33 seconds to spawn the instance on the hypervisor. [ 968.627481] env[62460]: DEBUG nova.compute.manager [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 968.627604] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40345a11-bfad-4ef6-831c-ba9931d16827 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.637549] env[62460]: DEBUG nova.scheduler.client.report [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Updating ProviderTree inventory for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 968.637762] env[62460]: DEBUG nova.compute.provider_tree [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 968.653293] env[62460]: DEBUG nova.scheduler.client.report [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Refreshing aggregate associations for resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728, aggregates: None {{(pid=62460) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 968.681468] env[62460]: DEBUG nova.scheduler.client.report [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Refreshing trait associations for resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62460) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 968.749832] env[62460]: DEBUG oslo_vmware.api [None req-daf435a1-8ec6-40b0-b5b0-b170e8f02000 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313962, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.816471] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24896598-9be6-4fce-abd3-8cbadd2d734e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.824256] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a3fad05-7d13-49d7-a6e9-450b5a176467 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.857538] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5155e9c9-66f7-4e17-bc9e-7aad2933487f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.868344] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6676d0f-f146-4072-b14c-3ffc77b05ddf {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.876026] env[62460]: DEBUG oslo_vmware.api [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313963, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.987389} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.876847] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 0d39a304-6a49-4d87-bfa2-4ba3b383578a/0d39a304-6a49-4d87-bfa2-4ba3b383578a.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 968.877091] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 968.877562] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6df7ffd6-3dee-4c5b-8b31-aa62426f5002 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.887339] env[62460]: DEBUG nova.compute.provider_tree [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 968.893856] env[62460]: DEBUG oslo_vmware.api [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 968.893856] env[62460]: value = "task-1313964" [ 968.893856] env[62460]: _type = "Task" [ 968.893856] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.903874] env[62460]: DEBUG oslo_vmware.api [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313964, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.149622] env[62460]: INFO nova.compute.manager [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Took 15.49 seconds to build instance. [ 969.245306] env[62460]: DEBUG oslo_vmware.api [None req-daf435a1-8ec6-40b0-b5b0-b170e8f02000 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313962, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.406368] env[62460]: DEBUG oslo_vmware.api [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313964, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.270497} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.406727] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 969.408028] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87cc6fc8-0e8a-497d-ac44-92da15edaed1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.431950] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Reconfiguring VM instance instance-00000060 to attach disk [datastore1] 0d39a304-6a49-4d87-bfa2-4ba3b383578a/0d39a304-6a49-4d87-bfa2-4ba3b383578a.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 969.433089] env[62460]: DEBUG nova.scheduler.client.report [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Updated inventory for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with generation 123 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 969.433290] env[62460]: DEBUG nova.compute.provider_tree [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Updating resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 generation from 123 to 124 during operation: update_inventory {{(pid=62460) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 969.433476] env[62460]: DEBUG nova.compute.provider_tree [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 969.438028] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-663198a2-7ef6-4af6-98cc-c7c5178458df {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.456172] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.616s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.456380] env[62460]: INFO nova.compute.manager [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Migrating [ 969.463324] env[62460]: DEBUG oslo_concurrency.lockutils [None req-189c95d3-7540-4455-870e-38113115b367 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.014s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.463559] env[62460]: DEBUG nova.objects.instance [None req-189c95d3-7540-4455-870e-38113115b367 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Lazy-loading 'resources' on Instance uuid bf28fc98-d9a9-4678-a107-ace2b6503353 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 969.476751] env[62460]: DEBUG oslo_vmware.api [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 969.476751] env[62460]: value = "task-1313965" [ 969.476751] env[62460]: _type = "Task" [ 969.476751] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.486671] env[62460]: DEBUG oslo_vmware.api [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313965, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.655267] env[62460]: DEBUG oslo_concurrency.lockutils [None req-5d938e04-8433-46a7-9259-6eb31ca57c96 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "f0f66557-04ef-4a96-8a56-e640bfda3b36" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.002s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.747240] env[62460]: DEBUG oslo_vmware.api [None req-daf435a1-8ec6-40b0-b5b0-b170e8f02000 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313962, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.871630] env[62460]: DEBUG oslo_concurrency.lockutils [None req-deacdf8b-d4e4-47fd-a4d8-402b1126291b tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "f0f66557-04ef-4a96-8a56-e640bfda3b36" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.871934] env[62460]: DEBUG oslo_concurrency.lockutils [None req-deacdf8b-d4e4-47fd-a4d8-402b1126291b tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "f0f66557-04ef-4a96-8a56-e640bfda3b36" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.872193] env[62460]: DEBUG oslo_concurrency.lockutils [None req-deacdf8b-d4e4-47fd-a4d8-402b1126291b tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "f0f66557-04ef-4a96-8a56-e640bfda3b36-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.873029] env[62460]: DEBUG oslo_concurrency.lockutils [None req-deacdf8b-d4e4-47fd-a4d8-402b1126291b tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "f0f66557-04ef-4a96-8a56-e640bfda3b36-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.873029] env[62460]: DEBUG oslo_concurrency.lockutils [None req-deacdf8b-d4e4-47fd-a4d8-402b1126291b tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "f0f66557-04ef-4a96-8a56-e640bfda3b36-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.876163] env[62460]: INFO nova.compute.manager [None req-deacdf8b-d4e4-47fd-a4d8-402b1126291b tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Terminating instance [ 969.878166] env[62460]: DEBUG nova.compute.manager [None req-deacdf8b-d4e4-47fd-a4d8-402b1126291b tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 969.878381] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-deacdf8b-d4e4-47fd-a4d8-402b1126291b tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 969.879468] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-994df0bd-bcc5-4cc8-9943-2d1be388e4ba {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.887239] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-deacdf8b-d4e4-47fd-a4d8-402b1126291b tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 969.887499] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-790be5b9-62fa-422b-bd4e-f59f4151f3e4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.894362] env[62460]: DEBUG oslo_vmware.api [None req-deacdf8b-d4e4-47fd-a4d8-402b1126291b tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for the task: (returnval){ [ 969.894362] env[62460]: value = "task-1313966" [ 969.894362] env[62460]: _type = "Task" [ 969.894362] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.909824] env[62460]: DEBUG oslo_vmware.api [None req-deacdf8b-d4e4-47fd-a4d8-402b1126291b tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313966, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.974211] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "refresh_cache-290500b5-d5b4-43b6-a843-20dacf561f4b" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 969.974517] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquired lock "refresh_cache-290500b5-d5b4-43b6-a843-20dacf561f4b" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.974751] env[62460]: DEBUG nova.network.neutron [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 969.992497] env[62460]: DEBUG oslo_vmware.api [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313965, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.122056] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86d10d6b-e92e-43ed-a97e-9db665573543 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.132071] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8fe5865-806a-4eb3-82b3-e47bd957c8a5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.163641] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50b640c1-e0d1-4a6f-8832-35731282924a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.171860] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1b7407c-122b-48bf-993a-b30db22d4961 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.185792] env[62460]: DEBUG nova.compute.provider_tree [None req-189c95d3-7540-4455-870e-38113115b367 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 970.247663] env[62460]: DEBUG oslo_vmware.api [None req-daf435a1-8ec6-40b0-b5b0-b170e8f02000 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313962, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.403964] env[62460]: DEBUG oslo_vmware.api [None req-deacdf8b-d4e4-47fd-a4d8-402b1126291b tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313966, 'name': PowerOffVM_Task, 'duration_secs': 0.192178} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.404301] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-deacdf8b-d4e4-47fd-a4d8-402b1126291b tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 970.404435] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-deacdf8b-d4e4-47fd-a4d8-402b1126291b tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 970.404698] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9a1617dc-f903-4633-bf6f-f49d01d88ae5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.491997] env[62460]: DEBUG oslo_vmware.api [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313965, 'name': ReconfigVM_Task, 'duration_secs': 0.658273} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.492285] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Reconfigured VM instance instance-00000060 to attach disk [datastore1] 0d39a304-6a49-4d87-bfa2-4ba3b383578a/0d39a304-6a49-4d87-bfa2-4ba3b383578a.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 970.492791] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8ba8aaca-b4c2-4058-b43d-4d6b1d5eccce {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.499144] env[62460]: DEBUG oslo_vmware.api [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 970.499144] env[62460]: value = "task-1313968" [ 970.499144] env[62460]: _type = "Task" [ 970.499144] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.507699] env[62460]: DEBUG oslo_vmware.api [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313968, 'name': Rename_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.513501] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-deacdf8b-d4e4-47fd-a4d8-402b1126291b tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 970.513501] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-deacdf8b-d4e4-47fd-a4d8-402b1126291b tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Deleting contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 970.513501] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-deacdf8b-d4e4-47fd-a4d8-402b1126291b tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Deleting the datastore file [datastore1] f0f66557-04ef-4a96-8a56-e640bfda3b36 {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 970.514135] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c2cbc36f-d589-4f30-aa35-b451ed0041af {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.522325] env[62460]: DEBUG oslo_vmware.api [None req-deacdf8b-d4e4-47fd-a4d8-402b1126291b tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for the task: (returnval){ [ 970.522325] env[62460]: value = "task-1313969" [ 970.522325] env[62460]: _type = "Task" [ 970.522325] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.532918] env[62460]: DEBUG oslo_vmware.api [None req-deacdf8b-d4e4-47fd-a4d8-402b1126291b tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313969, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.689532] env[62460]: DEBUG nova.scheduler.client.report [None req-189c95d3-7540-4455-870e-38113115b367 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 970.748739] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Acquiring lock "c772ecef-5cb5-4e69-a67c-0c487da4f1e3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.748981] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Lock "c772ecef-5cb5-4e69-a67c-0c487da4f1e3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.760019] env[62460]: DEBUG oslo_vmware.api [None req-daf435a1-8ec6-40b0-b5b0-b170e8f02000 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313962, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.789452] env[62460]: DEBUG nova.network.neutron [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Updating instance_info_cache with network_info: [{"id": "9f9c5b7b-922a-4387-b997-3f552dd85a50", "address": "fa:16:3e:78:80:d5", "network": {"id": "de789887-9cf9-4614-a7e1-dab430aaabaf", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-656400207-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0da056d93bdf40c39d6e82e457727ff6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5116f690-f825-4fee-8a47-42b073e716c5", "external-id": "nsx-vlan-transportzone-692", "segmentation_id": 692, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f9c5b7b-92", "ovs_interfaceid": "9f9c5b7b-922a-4387-b997-3f552dd85a50", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.009538] env[62460]: DEBUG oslo_vmware.api [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313968, 'name': Rename_Task, 'duration_secs': 0.146944} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.009817] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 971.010131] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-203dbb78-4f14-4f7f-8627-8a737c79df26 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.016632] env[62460]: DEBUG oslo_vmware.api [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 971.016632] env[62460]: value = "task-1313970" [ 971.016632] env[62460]: _type = "Task" [ 971.016632] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.025063] env[62460]: DEBUG oslo_vmware.api [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313970, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.036396] env[62460]: DEBUG oslo_vmware.api [None req-deacdf8b-d4e4-47fd-a4d8-402b1126291b tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313969, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.170962} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.036696] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-deacdf8b-d4e4-47fd-a4d8-402b1126291b tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 971.036905] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-deacdf8b-d4e4-47fd-a4d8-402b1126291b tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Deleted contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 971.037118] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-deacdf8b-d4e4-47fd-a4d8-402b1126291b tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 971.038172] env[62460]: INFO nova.compute.manager [None req-deacdf8b-d4e4-47fd-a4d8-402b1126291b tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Took 1.16 seconds to destroy the instance on the hypervisor. [ 971.038172] env[62460]: DEBUG oslo.service.loopingcall [None req-deacdf8b-d4e4-47fd-a4d8-402b1126291b tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 971.038172] env[62460]: DEBUG nova.compute.manager [-] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 971.038172] env[62460]: DEBUG nova.network.neutron [-] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 971.194666] env[62460]: DEBUG oslo_concurrency.lockutils [None req-189c95d3-7540-4455-870e-38113115b367 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.731s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.219866] env[62460]: INFO nova.scheduler.client.report [None req-189c95d3-7540-4455-870e-38113115b367 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Deleted allocations for instance bf28fc98-d9a9-4678-a107-ace2b6503353 [ 971.258552] env[62460]: DEBUG nova.compute.manager [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 971.261931] env[62460]: DEBUG oslo_vmware.api [None req-daf435a1-8ec6-40b0-b5b0-b170e8f02000 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313962, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.291941] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Releasing lock "refresh_cache-290500b5-d5b4-43b6-a843-20dacf561f4b" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.517342] env[62460]: DEBUG nova.compute.manager [req-07eccaa8-0902-4618-98a7-eadbeca2ae11 req-58341738-527c-48f0-9f84-ee5e0c892d38 service nova] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Received event network-vif-deleted-352f5563-2248-439c-a843-2001a2d5d230 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 971.517342] env[62460]: INFO nova.compute.manager [req-07eccaa8-0902-4618-98a7-eadbeca2ae11 req-58341738-527c-48f0-9f84-ee5e0c892d38 service nova] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Neutron deleted interface 352f5563-2248-439c-a843-2001a2d5d230; detaching it from the instance and deleting it from the info cache [ 971.517342] env[62460]: DEBUG nova.network.neutron [req-07eccaa8-0902-4618-98a7-eadbeca2ae11 req-58341738-527c-48f0-9f84-ee5e0c892d38 service nova] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.530866] env[62460]: DEBUG oslo_vmware.api [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313970, 'name': PowerOnVM_Task, 'duration_secs': 0.449882} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.531395] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 971.531794] env[62460]: INFO nova.compute.manager [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Took 8.91 seconds to spawn the instance on the hypervisor. [ 971.531886] env[62460]: DEBUG nova.compute.manager [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 971.532667] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6180c8d4-6819-4cd6-ab34-eb6e88e66680 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.734044] env[62460]: DEBUG oslo_concurrency.lockutils [None req-189c95d3-7540-4455-870e-38113115b367 tempest-AttachInterfacesUnderV243Test-414676011 tempest-AttachInterfacesUnderV243Test-414676011-project-member] Lock "bf28fc98-d9a9-4678-a107-ace2b6503353" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.655s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.758649] env[62460]: DEBUG oslo_vmware.api [None req-daf435a1-8ec6-40b0-b5b0-b170e8f02000 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313962, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.781043] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.781309] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.782816] env[62460]: INFO nova.compute.claims [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 971.990015] env[62460]: DEBUG nova.network.neutron [-] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 972.004873] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 972.005215] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 972.021224] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ac349f66-1697-4695-8b8a-b686a97e1d2e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.029918] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33ba0ed0-b5f8-4c01-95e4-f189d5f77939 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.057614] env[62460]: DEBUG nova.compute.manager [req-07eccaa8-0902-4618-98a7-eadbeca2ae11 req-58341738-527c-48f0-9f84-ee5e0c892d38 service nova] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Detach interface failed, port_id=352f5563-2248-439c-a843-2001a2d5d230, reason: Instance f0f66557-04ef-4a96-8a56-e640bfda3b36 could not be found. {{(pid=62460) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 972.058401] env[62460]: INFO nova.compute.manager [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Took 18.00 seconds to build instance. [ 972.259018] env[62460]: DEBUG oslo_vmware.api [None req-daf435a1-8ec6-40b0-b5b0-b170e8f02000 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313962, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.494409] env[62460]: INFO nova.compute.manager [-] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Took 1.46 seconds to deallocate network for instance. [ 972.511515] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 972.511757] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Starting heal instance info cache {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 972.560555] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3062b0ff-ec63-495e-b9a7-95b76bd195c9 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "0d39a304-6a49-4d87-bfa2-4ba3b383578a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.519s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.758828] env[62460]: DEBUG oslo_vmware.api [None req-daf435a1-8ec6-40b0-b5b0-b170e8f02000 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313962, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.818066] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-701983c6-f865-4955-a1e6-67c5c302cfc8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.839470] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Updating instance '290500b5-d5b4-43b6-a843-20dacf561f4b' progress to 0 {{(pid=62460) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 972.987404] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a7872dc-b7c1-4a9e-8689-3e5f3d49ea6a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.994413] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d92cfcde-2843-42d1-be71-89c7ed9c1cb9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.001107] env[62460]: DEBUG oslo_concurrency.lockutils [None req-deacdf8b-d4e4-47fd-a4d8-402b1126291b tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.056173] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Didn't find any instances for network info cache update. {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 973.058112] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 973.059251] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3c873e7-2bd7-40b3-a0e5-829e4e92279c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.062011] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 973.062538] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 973.063515] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 973.063791] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 973.063971] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 973.064270] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62460) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 973.064270] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 973.070110] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd9382df-a815-47de-af7d-3153b74be71f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.085368] env[62460]: DEBUG nova.compute.provider_tree [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 973.260206] env[62460]: DEBUG oslo_vmware.api [None req-daf435a1-8ec6-40b0-b5b0-b170e8f02000 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313962, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.345513] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 973.345885] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-32f22092-9059-47a5-9495-56f471dadc82 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.352960] env[62460]: DEBUG oslo_vmware.api [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 973.352960] env[62460]: value = "task-1313971" [ 973.352960] env[62460]: _type = "Task" [ 973.352960] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.362577] env[62460]: DEBUG oslo_vmware.api [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313971, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.521597] env[62460]: DEBUG nova.compute.manager [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Stashing vm_state: active {{(pid=62460) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 973.567528] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.589648] env[62460]: DEBUG nova.scheduler.client.report [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 973.761084] env[62460]: DEBUG oslo_vmware.api [None req-daf435a1-8ec6-40b0-b5b0-b170e8f02000 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313962, 'name': ReconfigVM_Task, 'duration_secs': 5.758107} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.761364] env[62460]: DEBUG oslo_concurrency.lockutils [None req-daf435a1-8ec6-40b0-b5b0-b170e8f02000 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Releasing lock "abd4501b-3ad8-4103-921d-5b80fc2f313f" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.761646] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-daf435a1-8ec6-40b0-b5b0-b170e8f02000 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Reconfigured VM to detach interface {{(pid=62460) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 973.862647] env[62460]: DEBUG oslo_vmware.api [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313971, 'name': PowerOffVM_Task, 'duration_secs': 0.175524} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.862932] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 973.863223] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Updating instance '290500b5-d5b4-43b6-a843-20dacf561f4b' progress to 17 {{(pid=62460) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 973.890059] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9716dc9e-a7e0-4ee0-bcee-e7f257dc8974 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Acquiring lock "5b6e8205-003e-49c0-a73d-be2e032a8272" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.890393] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9716dc9e-a7e0-4ee0-bcee-e7f257dc8974 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lock "5b6e8205-003e-49c0-a73d-be2e032a8272" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.890916] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9716dc9e-a7e0-4ee0-bcee-e7f257dc8974 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Acquiring lock "5b6e8205-003e-49c0-a73d-be2e032a8272-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.890916] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9716dc9e-a7e0-4ee0-bcee-e7f257dc8974 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lock "5b6e8205-003e-49c0-a73d-be2e032a8272-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.891097] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9716dc9e-a7e0-4ee0-bcee-e7f257dc8974 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lock "5b6e8205-003e-49c0-a73d-be2e032a8272-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.893160] env[62460]: INFO nova.compute.manager [None req-9716dc9e-a7e0-4ee0-bcee-e7f257dc8974 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Terminating instance [ 973.895117] env[62460]: DEBUG nova.compute.manager [None req-9716dc9e-a7e0-4ee0-bcee-e7f257dc8974 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 973.895227] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-9716dc9e-a7e0-4ee0-bcee-e7f257dc8974 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 973.896936] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-840d1ca9-0e49-495b-a0cb-a32fa5b9599d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.913898] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-9716dc9e-a7e0-4ee0-bcee-e7f257dc8974 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 973.914939] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4a92ae31-edad-4574-850a-9263a43a2ce6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.923773] env[62460]: DEBUG oslo_vmware.api [None req-9716dc9e-a7e0-4ee0-bcee-e7f257dc8974 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Waiting for the task: (returnval){ [ 973.923773] env[62460]: value = "task-1313972" [ 973.923773] env[62460]: _type = "Task" [ 973.923773] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.933810] env[62460]: DEBUG oslo_vmware.api [None req-9716dc9e-a7e0-4ee0-bcee-e7f257dc8974 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1313972, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.046510] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.095174] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.314s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.096306] env[62460]: DEBUG nova.compute.manager [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 974.099972] env[62460]: DEBUG oslo_concurrency.lockutils [None req-deacdf8b-d4e4-47fd-a4d8-402b1126291b tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.099s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.100422] env[62460]: DEBUG nova.objects.instance [None req-deacdf8b-d4e4-47fd-a4d8-402b1126291b tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lazy-loading 'resources' on Instance uuid f0f66557-04ef-4a96-8a56-e640bfda3b36 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 974.370194] env[62460]: DEBUG nova.virt.hardware [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 974.370461] env[62460]: DEBUG nova.virt.hardware [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 974.370703] env[62460]: DEBUG nova.virt.hardware [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 974.370871] env[62460]: DEBUG nova.virt.hardware [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 974.370954] env[62460]: DEBUG nova.virt.hardware [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 974.372034] env[62460]: DEBUG nova.virt.hardware [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 974.372034] env[62460]: DEBUG nova.virt.hardware [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 974.372034] env[62460]: DEBUG nova.virt.hardware [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 974.372034] env[62460]: DEBUG nova.virt.hardware [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 974.372034] env[62460]: DEBUG nova.virt.hardware [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 974.372034] env[62460]: DEBUG nova.virt.hardware [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 974.376898] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-63045046-0c7d-4b32-b164-889c35b87a07 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.393427] env[62460]: DEBUG oslo_vmware.api [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 974.393427] env[62460]: value = "task-1313973" [ 974.393427] env[62460]: _type = "Task" [ 974.393427] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.404028] env[62460]: DEBUG oslo_vmware.api [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313973, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.433628] env[62460]: DEBUG oslo_vmware.api [None req-9716dc9e-a7e0-4ee0-bcee-e7f257dc8974 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1313972, 'name': PowerOffVM_Task, 'duration_secs': 0.208079} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.433905] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-9716dc9e-a7e0-4ee0-bcee-e7f257dc8974 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 974.434101] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-9716dc9e-a7e0-4ee0-bcee-e7f257dc8974 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 974.434345] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8fd4d948-c1e3-43a4-8aae-f3a78df22758 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.495953] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-9716dc9e-a7e0-4ee0-bcee-e7f257dc8974 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 974.496264] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-9716dc9e-a7e0-4ee0-bcee-e7f257dc8974 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Deleting contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 974.496470] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-9716dc9e-a7e0-4ee0-bcee-e7f257dc8974 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Deleting the datastore file [datastore1] 5b6e8205-003e-49c0-a73d-be2e032a8272 {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 974.496748] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6d4e9f24-5fb4-4f11-b178-3cd68ec46ad3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.504250] env[62460]: DEBUG oslo_vmware.api [None req-9716dc9e-a7e0-4ee0-bcee-e7f257dc8974 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Waiting for the task: (returnval){ [ 974.504250] env[62460]: value = "task-1313975" [ 974.504250] env[62460]: _type = "Task" [ 974.504250] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.512205] env[62460]: DEBUG oslo_vmware.api [None req-9716dc9e-a7e0-4ee0-bcee-e7f257dc8974 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1313975, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.603968] env[62460]: DEBUG nova.compute.utils [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 974.608210] env[62460]: DEBUG nova.compute.manager [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 974.608393] env[62460]: DEBUG nova.network.neutron [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 974.711524] env[62460]: DEBUG nova.policy [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '48595264dab94ee59943f794a65778a9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '52e8bf27d39b46818b3674ccdee9896f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 974.761266] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0935412-5398-4823-ad24-62933343371b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.768795] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdb36dae-5df6-422f-b80e-2367673fb266 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.800058] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa958845-94b1-4bec-98e1-7a6f2e9f6b44 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.806938] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d6f4db0-e815-45a0-ac4d-8e29d0b0fc8f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.820055] env[62460]: DEBUG nova.compute.provider_tree [None req-deacdf8b-d4e4-47fd-a4d8-402b1126291b tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 974.903753] env[62460]: DEBUG oslo_vmware.api [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313973, 'name': ReconfigVM_Task, 'duration_secs': 0.139901} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.903935] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Updating instance '290500b5-d5b4-43b6-a843-20dacf561f4b' progress to 33 {{(pid=62460) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 975.016707] env[62460]: DEBUG oslo_vmware.api [None req-9716dc9e-a7e0-4ee0-bcee-e7f257dc8974 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1313975, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.317972} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.016982] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-9716dc9e-a7e0-4ee0-bcee-e7f257dc8974 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 975.017188] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-9716dc9e-a7e0-4ee0-bcee-e7f257dc8974 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Deleted contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 975.017395] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-9716dc9e-a7e0-4ee0-bcee-e7f257dc8974 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 975.017599] env[62460]: INFO nova.compute.manager [None req-9716dc9e-a7e0-4ee0-bcee-e7f257dc8974 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Took 1.12 seconds to destroy the instance on the hypervisor. [ 975.018121] env[62460]: DEBUG oslo.service.loopingcall [None req-9716dc9e-a7e0-4ee0-bcee-e7f257dc8974 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 975.018121] env[62460]: DEBUG nova.compute.manager [-] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 975.018245] env[62460]: DEBUG nova.network.neutron [-] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 975.046312] env[62460]: DEBUG nova.network.neutron [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] Successfully created port: 460727fc-e34d-424f-9770-d8dbea3ed624 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 975.112877] env[62460]: DEBUG nova.compute.manager [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 975.146228] env[62460]: DEBUG oslo_concurrency.lockutils [None req-daf435a1-8ec6-40b0-b5b0-b170e8f02000 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "refresh_cache-abd4501b-3ad8-4103-921d-5b80fc2f313f" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 975.146228] env[62460]: DEBUG oslo_concurrency.lockutils [None req-daf435a1-8ec6-40b0-b5b0-b170e8f02000 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquired lock "refresh_cache-abd4501b-3ad8-4103-921d-5b80fc2f313f" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.146228] env[62460]: DEBUG nova.network.neutron [None req-daf435a1-8ec6-40b0-b5b0-b170e8f02000 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 975.327826] env[62460]: DEBUG nova.scheduler.client.report [None req-deacdf8b-d4e4-47fd-a4d8-402b1126291b tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 975.410243] env[62460]: DEBUG nova.virt.hardware [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 975.410813] env[62460]: DEBUG nova.virt.hardware [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 975.411140] env[62460]: DEBUG nova.virt.hardware [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 975.411439] env[62460]: DEBUG nova.virt.hardware [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 975.412130] env[62460]: DEBUG nova.virt.hardware [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 975.412433] env[62460]: DEBUG nova.virt.hardware [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 975.412727] env[62460]: DEBUG nova.virt.hardware [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 975.412964] env[62460]: DEBUG nova.virt.hardware [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 975.413226] env[62460]: DEBUG nova.virt.hardware [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 975.413465] env[62460]: DEBUG nova.virt.hardware [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 975.413719] env[62460]: DEBUG nova.virt.hardware [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 975.420189] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Reconfiguring VM instance instance-0000005e to detach disk 2000 {{(pid=62460) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 975.421268] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2754d0fa-1670-41d2-ae2e-518342e8dd30 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.441528] env[62460]: DEBUG oslo_vmware.api [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 975.441528] env[62460]: value = "task-1313976" [ 975.441528] env[62460]: _type = "Task" [ 975.441528] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.450565] env[62460]: DEBUG oslo_vmware.api [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313976, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.833963] env[62460]: DEBUG oslo_concurrency.lockutils [None req-deacdf8b-d4e4-47fd-a4d8-402b1126291b tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.734s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.836400] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 2.269s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.836566] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.836784] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62460) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 975.837165] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 1.791s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.839160] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48d6b5b4-74e6-4d23-bddd-e8e2d2ff011a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.848739] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeb0a1ee-c07c-42a6-96a3-10002961f0dd {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.869793] env[62460]: INFO nova.scheduler.client.report [None req-deacdf8b-d4e4-47fd-a4d8-402b1126291b tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Deleted allocations for instance f0f66557-04ef-4a96-8a56-e640bfda3b36 [ 975.873927] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4325d071-0775-49d7-a9df-f8c7f0d990c3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.885585] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dac37dd-e254-4d1a-99b1-cfdf23935587 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.918782] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180656MB free_disk=135GB free_vcpus=48 pci_devices=None {{(pid=62460) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 975.918948] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.920057] env[62460]: INFO nova.network.neutron [None req-daf435a1-8ec6-40b0-b5b0-b170e8f02000 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Port 50d6a0c4-3681-4c1a-b42f-2987ee7a3bfb from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 975.920404] env[62460]: DEBUG nova.network.neutron [None req-daf435a1-8ec6-40b0-b5b0-b170e8f02000 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Updating instance_info_cache with network_info: [{"id": "546b0347-51db-4a4a-99e9-4f3db96a36bd", "address": "fa:16:3e:54:9b:7b", "network": {"id": "d3136e32-ad55-4b73-835c-8fa5f0480767", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1511414968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53fa06008e9f43488362895e7a143700", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "054fcd1e-638e-425a-a1de-78cb188ae026", "external-id": "nsx-vlan-transportzone-658", "segmentation_id": 658, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap546b0347-51", "ovs_interfaceid": "546b0347-51db-4a4a-99e9-4f3db96a36bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 975.956629] env[62460]: DEBUG oslo_vmware.api [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313976, 'name': ReconfigVM_Task, 'duration_secs': 0.158885} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.956926] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Reconfigured VM instance instance-0000005e to detach disk 2000 {{(pid=62460) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 975.957722] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a5cd048-7ce9-4c3a-8dfe-f8f545ce4e56 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.981014] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Reconfiguring VM instance instance-0000005e to attach disk [datastore1] 290500b5-d5b4-43b6-a843-20dacf561f4b/290500b5-d5b4-43b6-a843-20dacf561f4b.vmdk or device None with type thin {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 975.981328] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a0fccc68-6c46-40fa-87e9-e7e692515746 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.999161] env[62460]: DEBUG oslo_vmware.api [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 975.999161] env[62460]: value = "task-1313977" [ 975.999161] env[62460]: _type = "Task" [ 975.999161] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.007085] env[62460]: DEBUG oslo_vmware.api [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313977, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.123373] env[62460]: DEBUG nova.compute.manager [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 976.127611] env[62460]: DEBUG nova.compute.manager [req-accc8469-bd97-4c36-af73-77a50b594f0d req-3bc64941-f666-40f5-9515-a3ace348af45 service nova] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Received event network-vif-deleted-3515dec2-bd6c-4f96-aff1-8dd32de6f7d6 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 976.127807] env[62460]: INFO nova.compute.manager [req-accc8469-bd97-4c36-af73-77a50b594f0d req-3bc64941-f666-40f5-9515-a3ace348af45 service nova] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Neutron deleted interface 3515dec2-bd6c-4f96-aff1-8dd32de6f7d6; detaching it from the instance and deleting it from the info cache [ 976.128382] env[62460]: DEBUG nova.network.neutron [req-accc8469-bd97-4c36-af73-77a50b594f0d req-3bc64941-f666-40f5-9515-a3ace348af45 service nova] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 976.152780] env[62460]: DEBUG nova.virt.hardware [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 976.152780] env[62460]: DEBUG nova.virt.hardware [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 976.152780] env[62460]: DEBUG nova.virt.hardware [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 976.152780] env[62460]: DEBUG nova.virt.hardware [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 976.152780] env[62460]: DEBUG nova.virt.hardware [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 976.152780] env[62460]: DEBUG nova.virt.hardware [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 976.153131] env[62460]: DEBUG nova.virt.hardware [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 976.153131] env[62460]: DEBUG nova.virt.hardware [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 976.153131] env[62460]: DEBUG nova.virt.hardware [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 976.153303] env[62460]: DEBUG nova.virt.hardware [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 976.153483] env[62460]: DEBUG nova.virt.hardware [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 976.154661] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f76cb59-3d09-42ec-a963-9f936e0a3f80 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.163575] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9538fdc-68cd-42db-89fc-67a494a7d9d7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.345104] env[62460]: INFO nova.compute.claims [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 976.384374] env[62460]: DEBUG oslo_concurrency.lockutils [None req-deacdf8b-d4e4-47fd-a4d8-402b1126291b tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "f0f66557-04ef-4a96-8a56-e640bfda3b36" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.512s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.423560] env[62460]: DEBUG oslo_concurrency.lockutils [None req-daf435a1-8ec6-40b0-b5b0-b170e8f02000 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Releasing lock "refresh_cache-abd4501b-3ad8-4103-921d-5b80fc2f313f" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 976.509175] env[62460]: DEBUG oslo_vmware.api [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313977, 'name': ReconfigVM_Task, 'duration_secs': 0.296566} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.509547] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Reconfigured VM instance instance-0000005e to attach disk [datastore1] 290500b5-d5b4-43b6-a843-20dacf561f4b/290500b5-d5b4-43b6-a843-20dacf561f4b.vmdk or device None with type thin {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 976.509889] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Updating instance '290500b5-d5b4-43b6-a843-20dacf561f4b' progress to 50 {{(pid=62460) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 976.600256] env[62460]: DEBUG nova.network.neutron [-] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 976.630997] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bb2dff8d-1f7e-4818-b487-ee5d8bee6801 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.641150] env[62460]: DEBUG nova.compute.manager [req-f99ea9c0-2410-44e0-826e-df30225c74ba req-6aff97e2-74c3-4c45-ab2e-c03a8fb07453 service nova] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] Received event network-vif-plugged-460727fc-e34d-424f-9770-d8dbea3ed624 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 976.641150] env[62460]: DEBUG oslo_concurrency.lockutils [req-f99ea9c0-2410-44e0-826e-df30225c74ba req-6aff97e2-74c3-4c45-ab2e-c03a8fb07453 service nova] Acquiring lock "c772ecef-5cb5-4e69-a67c-0c487da4f1e3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.641323] env[62460]: DEBUG oslo_concurrency.lockutils [req-f99ea9c0-2410-44e0-826e-df30225c74ba req-6aff97e2-74c3-4c45-ab2e-c03a8fb07453 service nova] Lock "c772ecef-5cb5-4e69-a67c-0c487da4f1e3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.642045] env[62460]: DEBUG oslo_concurrency.lockutils [req-f99ea9c0-2410-44e0-826e-df30225c74ba req-6aff97e2-74c3-4c45-ab2e-c03a8fb07453 service nova] Lock "c772ecef-5cb5-4e69-a67c-0c487da4f1e3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.642045] env[62460]: DEBUG nova.compute.manager [req-f99ea9c0-2410-44e0-826e-df30225c74ba req-6aff97e2-74c3-4c45-ab2e-c03a8fb07453 service nova] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] No waiting events found dispatching network-vif-plugged-460727fc-e34d-424f-9770-d8dbea3ed624 {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 976.642045] env[62460]: WARNING nova.compute.manager [req-f99ea9c0-2410-44e0-826e-df30225c74ba req-6aff97e2-74c3-4c45-ab2e-c03a8fb07453 service nova] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] Received unexpected event network-vif-plugged-460727fc-e34d-424f-9770-d8dbea3ed624 for instance with vm_state building and task_state spawning. [ 976.646939] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5dfe53f-9b34-4b81-bb11-99703663a168 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.675354] env[62460]: DEBUG nova.compute.manager [req-accc8469-bd97-4c36-af73-77a50b594f0d req-3bc64941-f666-40f5-9515-a3ace348af45 service nova] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Detach interface failed, port_id=3515dec2-bd6c-4f96-aff1-8dd32de6f7d6, reason: Instance 5b6e8205-003e-49c0-a73d-be2e032a8272 could not be found. {{(pid=62460) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 976.735481] env[62460]: DEBUG nova.network.neutron [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] Successfully updated port: 460727fc-e34d-424f-9770-d8dbea3ed624 {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 976.852620] env[62460]: INFO nova.compute.resource_tracker [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Updating resource usage from migration f48aedf1-a18c-4d31-9f53-6e452541bb4d [ 976.927886] env[62460]: DEBUG oslo_concurrency.lockutils [None req-daf435a1-8ec6-40b0-b5b0-b170e8f02000 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "interface-abd4501b-3ad8-4103-921d-5b80fc2f313f-50d6a0c4-3681-4c1a-b42f-2987ee7a3bfb" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.771s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.017757] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-343d64ba-3616-4098-a45d-9b0e4306c764 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.023253] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2509c712-9396-4345-91a7-eef3326f1b04 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.051336] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e12c0d4e-1066-4b06-bdc4-80107ef7b744 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.053229] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95c702b6-cb7f-4e21-98e0-1d0bc70ec52b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.070643] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Updating instance '290500b5-d5b4-43b6-a843-20dacf561f4b' progress to 67 {{(pid=62460) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 977.100577] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e9e6d6a-4bdd-44f9-8032-e4bbb75b0605 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.103412] env[62460]: INFO nova.compute.manager [-] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Took 2.09 seconds to deallocate network for instance. [ 977.111013] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e83c354-8b0f-4487-92c8-c5fbb52b0e7d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.126177] env[62460]: DEBUG nova.compute.provider_tree [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 977.238467] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Acquiring lock "refresh_cache-c772ecef-5cb5-4e69-a67c-0c487da4f1e3" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.238734] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Acquired lock "refresh_cache-c772ecef-5cb5-4e69-a67c-0c487da4f1e3" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.238805] env[62460]: DEBUG nova.network.neutron [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 977.612264] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9716dc9e-a7e0-4ee0-bcee-e7f257dc8974 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.629823] env[62460]: DEBUG nova.scheduler.client.report [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 977.664013] env[62460]: DEBUG nova.network.neutron [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Port 9f9c5b7b-922a-4387-b997-3f552dd85a50 binding to destination host cpu-1 is already ACTIVE {{(pid=62460) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 977.665609] env[62460]: DEBUG oslo_concurrency.lockutils [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "4c5e60fe-6366-4fb9-975f-52345de67478" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.667076] env[62460]: DEBUG oslo_concurrency.lockutils [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "4c5e60fe-6366-4fb9-975f-52345de67478" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.779480] env[62460]: DEBUG nova.network.neutron [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 977.930625] env[62460]: DEBUG nova.network.neutron [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] Updating instance_info_cache with network_info: [{"id": "460727fc-e34d-424f-9770-d8dbea3ed624", "address": "fa:16:3e:2a:c1:c6", "network": {"id": "23f21687-24fb-42e7-83be-8001ab8d5ab6", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-573166759-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52e8bf27d39b46818b3674ccdee9896f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "41f66e20-fd86-4158-bbdc-7a150e85e844", "external-id": "nsx-vlan-transportzone-182", "segmentation_id": 182, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap460727fc-e3", "ovs_interfaceid": "460727fc-e34d-424f-9770-d8dbea3ed624", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 978.138044] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.300s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.138471] env[62460]: INFO nova.compute.manager [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Migrating [ 978.145233] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 2.226s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.164274] env[62460]: DEBUG nova.compute.manager [req-2f252607-c1c7-4b42-bdf4-5f9dd6269ac1 req-4e8a11c8-a3c1-41df-b73c-844cd3c733fd service nova] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Received event network-changed-546b0347-51db-4a4a-99e9-4f3db96a36bd {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 978.164274] env[62460]: DEBUG nova.compute.manager [req-2f252607-c1c7-4b42-bdf4-5f9dd6269ac1 req-4e8a11c8-a3c1-41df-b73c-844cd3c733fd service nova] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Refreshing instance network info cache due to event network-changed-546b0347-51db-4a4a-99e9-4f3db96a36bd. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 978.164274] env[62460]: DEBUG oslo_concurrency.lockutils [req-2f252607-c1c7-4b42-bdf4-5f9dd6269ac1 req-4e8a11c8-a3c1-41df-b73c-844cd3c733fd service nova] Acquiring lock "refresh_cache-abd4501b-3ad8-4103-921d-5b80fc2f313f" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.164274] env[62460]: DEBUG oslo_concurrency.lockutils [req-2f252607-c1c7-4b42-bdf4-5f9dd6269ac1 req-4e8a11c8-a3c1-41df-b73c-844cd3c733fd service nova] Acquired lock "refresh_cache-abd4501b-3ad8-4103-921d-5b80fc2f313f" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.164274] env[62460]: DEBUG nova.network.neutron [req-2f252607-c1c7-4b42-bdf4-5f9dd6269ac1 req-4e8a11c8-a3c1-41df-b73c-844cd3c733fd service nova] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Refreshing network info cache for port 546b0347-51db-4a4a-99e9-4f3db96a36bd {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 978.173781] env[62460]: DEBUG nova.compute.manager [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 978.434055] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Releasing lock "refresh_cache-c772ecef-5cb5-4e69-a67c-0c487da4f1e3" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.434350] env[62460]: DEBUG nova.compute.manager [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] Instance network_info: |[{"id": "460727fc-e34d-424f-9770-d8dbea3ed624", "address": "fa:16:3e:2a:c1:c6", "network": {"id": "23f21687-24fb-42e7-83be-8001ab8d5ab6", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-573166759-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52e8bf27d39b46818b3674ccdee9896f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "41f66e20-fd86-4158-bbdc-7a150e85e844", "external-id": "nsx-vlan-transportzone-182", "segmentation_id": 182, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap460727fc-e3", "ovs_interfaceid": "460727fc-e34d-424f-9770-d8dbea3ed624", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 978.434840] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2a:c1:c6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '41f66e20-fd86-4158-bbdc-7a150e85e844', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '460727fc-e34d-424f-9770-d8dbea3ed624', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 978.442505] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Creating folder: Project (52e8bf27d39b46818b3674ccdee9896f). Parent ref: group-v281134. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 978.442796] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-37b7471d-f430-4799-8460-0973e9de4431 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.457329] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Created folder: Project (52e8bf27d39b46818b3674ccdee9896f) in parent group-v281134. [ 978.457450] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Creating folder: Instances. Parent ref: group-v281257. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 978.457711] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8f74c8b5-cb99-4be1-9b46-7d2c434a6d2f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.468111] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Created folder: Instances in parent group-v281257. [ 978.468363] env[62460]: DEBUG oslo.service.loopingcall [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 978.468565] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 978.468778] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fe1a3278-92c2-4aa6-a312-218174755303 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.488368] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 978.488368] env[62460]: value = "task-1313980" [ 978.488368] env[62460]: _type = "Task" [ 978.488368] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.498012] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313980, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.660467] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "refresh_cache-0d39a304-6a49-4d87-bfa2-4ba3b383578a" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.660533] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquired lock "refresh_cache-0d39a304-6a49-4d87-bfa2-4ba3b383578a" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.660982] env[62460]: DEBUG nova.network.neutron [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 978.674720] env[62460]: DEBUG nova.compute.manager [req-17350fad-90a3-4ff2-9e6c-ab8990528a4d req-afb2ddc8-b9bf-40f5-abd3-0e1153acecc1 service nova] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] Received event network-changed-460727fc-e34d-424f-9770-d8dbea3ed624 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 978.675148] env[62460]: DEBUG nova.compute.manager [req-17350fad-90a3-4ff2-9e6c-ab8990528a4d req-afb2ddc8-b9bf-40f5-abd3-0e1153acecc1 service nova] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] Refreshing instance network info cache due to event network-changed-460727fc-e34d-424f-9770-d8dbea3ed624. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 978.675331] env[62460]: DEBUG oslo_concurrency.lockutils [req-17350fad-90a3-4ff2-9e6c-ab8990528a4d req-afb2ddc8-b9bf-40f5-abd3-0e1153acecc1 service nova] Acquiring lock "refresh_cache-c772ecef-5cb5-4e69-a67c-0c487da4f1e3" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.675484] env[62460]: DEBUG oslo_concurrency.lockutils [req-17350fad-90a3-4ff2-9e6c-ab8990528a4d req-afb2ddc8-b9bf-40f5-abd3-0e1153acecc1 service nova] Acquired lock "refresh_cache-c772ecef-5cb5-4e69-a67c-0c487da4f1e3" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.675658] env[62460]: DEBUG nova.network.neutron [req-17350fad-90a3-4ff2-9e6c-ab8990528a4d req-afb2ddc8-b9bf-40f5-abd3-0e1153acecc1 service nova] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] Refreshing network info cache for port 460727fc-e34d-424f-9770-d8dbea3ed624 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 978.700940] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "290500b5-d5b4-43b6-a843-20dacf561f4b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.701161] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "290500b5-d5b4-43b6-a843-20dacf561f4b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.701349] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "290500b5-d5b4-43b6-a843-20dacf561f4b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.728908] env[62460]: DEBUG oslo_concurrency.lockutils [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.995583] env[62460]: DEBUG nova.network.neutron [req-17350fad-90a3-4ff2-9e6c-ab8990528a4d req-afb2ddc8-b9bf-40f5-abd3-0e1153acecc1 service nova] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] Updated VIF entry in instance network info cache for port 460727fc-e34d-424f-9770-d8dbea3ed624. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 978.995583] env[62460]: DEBUG nova.network.neutron [req-17350fad-90a3-4ff2-9e6c-ab8990528a4d req-afb2ddc8-b9bf-40f5-abd3-0e1153acecc1 service nova] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] Updating instance_info_cache with network_info: [{"id": "460727fc-e34d-424f-9770-d8dbea3ed624", "address": "fa:16:3e:2a:c1:c6", "network": {"id": "23f21687-24fb-42e7-83be-8001ab8d5ab6", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-573166759-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52e8bf27d39b46818b3674ccdee9896f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "41f66e20-fd86-4158-bbdc-7a150e85e844", "external-id": "nsx-vlan-transportzone-182", "segmentation_id": 182, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap460727fc-e3", "ovs_interfaceid": "460727fc-e34d-424f-9770-d8dbea3ed624", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 979.000785] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313980, 'name': CreateVM_Task, 'duration_secs': 0.314098} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.001167] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 979.001968] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.002293] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.002727] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 979.003114] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7cfe8c97-e19d-4016-b736-af7b6bd90561 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.010130] env[62460]: DEBUG oslo_vmware.api [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Waiting for the task: (returnval){ [ 979.010130] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52c5980e-d6c5-0448-9c6a-8d23cb115a9a" [ 979.010130] env[62460]: _type = "Task" [ 979.010130] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.018098] env[62460]: DEBUG oslo_vmware.api [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52c5980e-d6c5-0448-9c6a-8d23cb115a9a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.019122] env[62460]: DEBUG nova.network.neutron [req-2f252607-c1c7-4b42-bdf4-5f9dd6269ac1 req-4e8a11c8-a3c1-41df-b73c-844cd3c733fd service nova] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Updated VIF entry in instance network info cache for port 546b0347-51db-4a4a-99e9-4f3db96a36bd. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 979.019611] env[62460]: DEBUG nova.network.neutron [req-2f252607-c1c7-4b42-bdf4-5f9dd6269ac1 req-4e8a11c8-a3c1-41df-b73c-844cd3c733fd service nova] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Updating instance_info_cache with network_info: [{"id": "546b0347-51db-4a4a-99e9-4f3db96a36bd", "address": "fa:16:3e:54:9b:7b", "network": {"id": "d3136e32-ad55-4b73-835c-8fa5f0480767", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1511414968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53fa06008e9f43488362895e7a143700", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "054fcd1e-638e-425a-a1de-78cb188ae026", "external-id": "nsx-vlan-transportzone-658", "segmentation_id": 658, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap546b0347-51", "ovs_interfaceid": "546b0347-51db-4a4a-99e9-4f3db96a36bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 979.163186] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Applying migration context for instance 0d39a304-6a49-4d87-bfa2-4ba3b383578a as it has an incoming, in-progress migration f48aedf1-a18c-4d31-9f53-6e452541bb4d. Migration status is pre-migrating {{(pid=62460) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 979.163867] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Applying migration context for instance 290500b5-d5b4-43b6-a843-20dacf561f4b as it has an incoming, in-progress migration 6c432a03-9e05-49cc-bf18-7f0bc8bde6eb. Migration status is post-migrating {{(pid=62460) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 979.164266] env[62460]: INFO nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Updating resource usage from migration f48aedf1-a18c-4d31-9f53-6e452541bb4d [ 979.165160] env[62460]: INFO nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Updating resource usage from migration 6c432a03-9e05-49cc-bf18-7f0bc8bde6eb [ 979.195724] env[62460]: WARNING nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 5b6e8205-003e-49c0-a73d-be2e032a8272 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 979.195894] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance abd4501b-3ad8-4103-921d-5b80fc2f313f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 979.196039] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 896f8ff4-e45f-4403-a727-03ee25e58609 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 979.196166] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 979.196284] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Migration 6c432a03-9e05-49cc-bf18-7f0bc8bde6eb is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 979.196401] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 290500b5-d5b4-43b6-a843-20dacf561f4b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 979.196517] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance c772ecef-5cb5-4e69-a67c-0c487da4f1e3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 979.196633] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Migration f48aedf1-a18c-4d31-9f53-6e452541bb4d is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 979.196747] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 0d39a304-6a49-4d87-bfa2-4ba3b383578a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 979.400903] env[62460]: DEBUG nova.network.neutron [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Updating instance_info_cache with network_info: [{"id": "786ef177-843f-44d0-8920-2f4332e29154", "address": "fa:16:3e:09:59:ab", "network": {"id": "3c49720c-94c9-4e2f-bbb7-4266cc97faee", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1650281790-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfa5a65b0f614e769de5b3aa77bf869e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap786ef177-84", "ovs_interfaceid": "786ef177-843f-44d0-8920-2f4332e29154", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 979.498107] env[62460]: DEBUG oslo_concurrency.lockutils [req-17350fad-90a3-4ff2-9e6c-ab8990528a4d req-afb2ddc8-b9bf-40f5-abd3-0e1153acecc1 service nova] Releasing lock "refresh_cache-c772ecef-5cb5-4e69-a67c-0c487da4f1e3" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.522027] env[62460]: DEBUG oslo_vmware.api [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52c5980e-d6c5-0448-9c6a-8d23cb115a9a, 'name': SearchDatastore_Task, 'duration_secs': 0.01052} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.522027] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.522027] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 979.522027] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.522027] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.522027] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 979.522027] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c2a05f78-c67e-4ac0-abc5-940862dc9dbd {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.522027] env[62460]: DEBUG oslo_concurrency.lockutils [req-2f252607-c1c7-4b42-bdf4-5f9dd6269ac1 req-4e8a11c8-a3c1-41df-b73c-844cd3c733fd service nova] Releasing lock "refresh_cache-abd4501b-3ad8-4103-921d-5b80fc2f313f" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.522637] env[62460]: DEBUG nova.compute.manager [req-2f252607-c1c7-4b42-bdf4-5f9dd6269ac1 req-4e8a11c8-a3c1-41df-b73c-844cd3c733fd service nova] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Received event network-changed-90a7d774-4d76-4525-8e6d-8c08fdb93fb4 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 979.522940] env[62460]: DEBUG nova.compute.manager [req-2f252607-c1c7-4b42-bdf4-5f9dd6269ac1 req-4e8a11c8-a3c1-41df-b73c-844cd3c733fd service nova] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Refreshing instance network info cache due to event network-changed-90a7d774-4d76-4525-8e6d-8c08fdb93fb4. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 979.523300] env[62460]: DEBUG oslo_concurrency.lockutils [req-2f252607-c1c7-4b42-bdf4-5f9dd6269ac1 req-4e8a11c8-a3c1-41df-b73c-844cd3c733fd service nova] Acquiring lock "refresh_cache-0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.523650] env[62460]: DEBUG oslo_concurrency.lockutils [req-2f252607-c1c7-4b42-bdf4-5f9dd6269ac1 req-4e8a11c8-a3c1-41df-b73c-844cd3c733fd service nova] Acquired lock "refresh_cache-0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.523976] env[62460]: DEBUG nova.network.neutron [req-2f252607-c1c7-4b42-bdf4-5f9dd6269ac1 req-4e8a11c8-a3c1-41df-b73c-844cd3c733fd service nova] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Refreshing network info cache for port 90a7d774-4d76-4525-8e6d-8c08fdb93fb4 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 979.531380] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 979.531558] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 979.532839] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ecdbf979-bcf8-4554-9906-0341722eeb4e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.540667] env[62460]: DEBUG oslo_vmware.api [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Waiting for the task: (returnval){ [ 979.540667] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52a94ee0-ea42-6bfc-440f-dd9c23b708f2" [ 979.540667] env[62460]: _type = "Task" [ 979.540667] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.274951] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 4c5e60fe-6366-4fb9-975f-52345de67478 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 980.275320] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=62460) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 980.275371] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2176MB phys_disk=200GB used_disk=8GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=62460) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 980.281276] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Releasing lock "refresh_cache-0d39a304-6a49-4d87-bfa2-4ba3b383578a" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.286023] env[62460]: DEBUG oslo_vmware.api [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52a94ee0-ea42-6bfc-440f-dd9c23b708f2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.286023] env[62460]: WARNING oslo_vmware.common.loopingcall [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] task run outlasted interval by 0.24364399999999997 sec [ 980.287990] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "61f4666e-0588-4a00-a969-2349148caef2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.288233] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "61f4666e-0588-4a00-a969-2349148caef2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.298980] env[62460]: DEBUG oslo_vmware.api [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52a94ee0-ea42-6bfc-440f-dd9c23b708f2, 'name': SearchDatastore_Task, 'duration_secs': 0.023732} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.300499] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6d3f31b8-bf35-43c0-96ed-49b6cb0aa843 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.308237] env[62460]: DEBUG oslo_vmware.api [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Waiting for the task: (returnval){ [ 980.308237] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]524b09b3-5f5b-7a0f-ce33-8dd5ac943279" [ 980.308237] env[62460]: _type = "Task" [ 980.308237] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.318468] env[62460]: DEBUG oslo_vmware.api [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]524b09b3-5f5b-7a0f-ce33-8dd5ac943279, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.326959] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "refresh_cache-290500b5-d5b4-43b6-a843-20dacf561f4b" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 980.327391] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquired lock "refresh_cache-290500b5-d5b4-43b6-a843-20dacf561f4b" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 980.327391] env[62460]: DEBUG nova.network.neutron [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 980.450796] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6afd0a3a-c4d6-4d5f-82ec-3498465c4aef {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.458368] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-760d3479-582e-490b-a1f3-a348b7ead441 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.489404] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59cef02f-3a45-4d56-afb7-1f3c7dff4d04 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.497112] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd68315f-c119-4a2d-886f-1d9d91657196 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.511104] env[62460]: DEBUG nova.compute.provider_tree [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 980.630478] env[62460]: DEBUG nova.network.neutron [req-2f252607-c1c7-4b42-bdf4-5f9dd6269ac1 req-4e8a11c8-a3c1-41df-b73c-844cd3c733fd service nova] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Updated VIF entry in instance network info cache for port 90a7d774-4d76-4525-8e6d-8c08fdb93fb4. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 980.631059] env[62460]: DEBUG nova.network.neutron [req-2f252607-c1c7-4b42-bdf4-5f9dd6269ac1 req-4e8a11c8-a3c1-41df-b73c-844cd3c733fd service nova] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Updating instance_info_cache with network_info: [{"id": "90a7d774-4d76-4525-8e6d-8c08fdb93fb4", "address": "fa:16:3e:5b:1c:b7", "network": {"id": "d3136e32-ad55-4b73-835c-8fa5f0480767", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1511414968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53fa06008e9f43488362895e7a143700", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "054fcd1e-638e-425a-a1de-78cb188ae026", "external-id": "nsx-vlan-transportzone-658", "segmentation_id": 658, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90a7d774-4d", "ovs_interfaceid": "90a7d774-4d76-4525-8e6d-8c08fdb93fb4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.795639] env[62460]: DEBUG nova.compute.manager [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 980.819967] env[62460]: DEBUG oslo_vmware.api [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]524b09b3-5f5b-7a0f-ce33-8dd5ac943279, 'name': SearchDatastore_Task, 'duration_secs': 0.009774} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.820155] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.821040] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] c772ecef-5cb5-4e69-a67c-0c487da4f1e3/c772ecef-5cb5-4e69-a67c-0c487da4f1e3.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 980.821040] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-aa979bad-e5c6-403e-a0e5-263ef00ca352 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.827015] env[62460]: DEBUG oslo_vmware.api [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Waiting for the task: (returnval){ [ 980.827015] env[62460]: value = "task-1313981" [ 980.827015] env[62460]: _type = "Task" [ 980.827015] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.836777] env[62460]: DEBUG oslo_vmware.api [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Task: {'id': task-1313981, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.014637] env[62460]: DEBUG nova.scheduler.client.report [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 981.066573] env[62460]: DEBUG nova.network.neutron [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Updating instance_info_cache with network_info: [{"id": "9f9c5b7b-922a-4387-b997-3f552dd85a50", "address": "fa:16:3e:78:80:d5", "network": {"id": "de789887-9cf9-4614-a7e1-dab430aaabaf", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-656400207-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0da056d93bdf40c39d6e82e457727ff6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5116f690-f825-4fee-8a47-42b073e716c5", "external-id": "nsx-vlan-transportzone-692", "segmentation_id": 692, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f9c5b7b-92", "ovs_interfaceid": "9f9c5b7b-922a-4387-b997-3f552dd85a50", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 981.134142] env[62460]: DEBUG oslo_concurrency.lockutils [req-2f252607-c1c7-4b42-bdf4-5f9dd6269ac1 req-4e8a11c8-a3c1-41df-b73c-844cd3c733fd service nova] Releasing lock "refresh_cache-0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 981.318119] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.336233] env[62460]: DEBUG oslo_vmware.api [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Task: {'id': task-1313981, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.460543} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.336527] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] c772ecef-5cb5-4e69-a67c-0c487da4f1e3/c772ecef-5cb5-4e69-a67c-0c487da4f1e3.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 981.336747] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 981.337028] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-50d71599-26b9-41a7-9e2e-3d3576f45cfe {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.344597] env[62460]: DEBUG oslo_vmware.api [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Waiting for the task: (returnval){ [ 981.344597] env[62460]: value = "task-1313982" [ 981.344597] env[62460]: _type = "Task" [ 981.344597] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.352641] env[62460]: DEBUG oslo_vmware.api [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Task: {'id': task-1313982, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.519848] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62460) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 981.520167] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.375s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.520749] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9716dc9e-a7e0-4ee0-bcee-e7f257dc8974 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.909s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.521074] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9716dc9e-a7e0-4ee0-bcee-e7f257dc8974 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.523294] env[62460]: DEBUG oslo_concurrency.lockutils [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.795s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.525362] env[62460]: INFO nova.compute.claims [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 981.549868] env[62460]: INFO nova.scheduler.client.report [None req-9716dc9e-a7e0-4ee0-bcee-e7f257dc8974 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Deleted allocations for instance 5b6e8205-003e-49c0-a73d-be2e032a8272 [ 981.569939] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Releasing lock "refresh_cache-290500b5-d5b4-43b6-a843-20dacf561f4b" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 981.799141] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3392f888-27bf-4ba5-922e-9a73fe58a604 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.819822] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Updating instance '0d39a304-6a49-4d87-bfa2-4ba3b383578a' progress to 0 {{(pid=62460) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 981.853765] env[62460]: DEBUG oslo_vmware.api [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Task: {'id': task-1313982, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.058351} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.854082] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 981.854936] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4cd427f-5f9f-450e-90ef-366e451878bf {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.876671] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] Reconfiguring VM instance instance-00000061 to attach disk [datastore2] c772ecef-5cb5-4e69-a67c-0c487da4f1e3/c772ecef-5cb5-4e69-a67c-0c487da4f1e3.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 981.877743] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0e15c073-b5b9-4e0c-8716-43ec8ab68298 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.896785] env[62460]: DEBUG oslo_vmware.api [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Waiting for the task: (returnval){ [ 981.896785] env[62460]: value = "task-1313983" [ 981.896785] env[62460]: _type = "Task" [ 981.896785] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.905811] env[62460]: DEBUG oslo_vmware.api [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Task: {'id': task-1313983, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.059390] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9716dc9e-a7e0-4ee0-bcee-e7f257dc8974 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lock "5b6e8205-003e-49c0-a73d-be2e032a8272" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.169s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.091442] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c93351a4-11fd-4c3a-ae11-02fa931cfa86 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.112840] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9fad177-8735-434a-a603-cacd362dc23e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.120314] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Updating instance '290500b5-d5b4-43b6-a843-20dacf561f4b' progress to 83 {{(pid=62460) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 982.326953] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 982.328023] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-47b8688b-8bc9-47de-ba80-2f77ba4fa1d8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.335099] env[62460]: DEBUG oslo_vmware.api [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 982.335099] env[62460]: value = "task-1313984" [ 982.335099] env[62460]: _type = "Task" [ 982.335099] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.343321] env[62460]: DEBUG oslo_vmware.api [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313984, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.406993] env[62460]: DEBUG oslo_vmware.api [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Task: {'id': task-1313983, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.628117] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 982.628735] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-46653f15-4937-4f6f-a69d-619d94ce5988 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.636047] env[62460]: DEBUG oslo_vmware.api [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 982.636047] env[62460]: value = "task-1313985" [ 982.636047] env[62460]: _type = "Task" [ 982.636047] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.647348] env[62460]: DEBUG oslo_vmware.api [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313985, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.676176] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fd153d9-535b-4eed-9588-5a75bc45a1ea {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.685452] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f818b201-e6d7-417c-946f-5d90a565a514 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.714556] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80eba826-c2a1-4b32-ae21-ce0287dab325 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.721732] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bf1e2a3-490d-4832-9b53-ea87afb81c83 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.737737] env[62460]: DEBUG nova.compute.provider_tree [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 982.845036] env[62460]: DEBUG oslo_vmware.api [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313984, 'name': PowerOffVM_Task, 'duration_secs': 0.27521} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.845345] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 982.845541] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Updating instance '0d39a304-6a49-4d87-bfa2-4ba3b383578a' progress to 17 {{(pid=62460) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 982.907238] env[62460]: DEBUG oslo_vmware.api [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Task: {'id': task-1313983, 'name': ReconfigVM_Task, 'duration_secs': 0.772808} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.907541] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] Reconfigured VM instance instance-00000061 to attach disk [datastore2] c772ecef-5cb5-4e69-a67c-0c487da4f1e3/c772ecef-5cb5-4e69-a67c-0c487da4f1e3.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 982.908176] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2118721c-7c07-40bf-a57d-15d57986c46d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.913876] env[62460]: DEBUG oslo_vmware.api [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Waiting for the task: (returnval){ [ 982.913876] env[62460]: value = "task-1313986" [ 982.913876] env[62460]: _type = "Task" [ 982.913876] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.921434] env[62460]: DEBUG oslo_vmware.api [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Task: {'id': task-1313986, 'name': Rename_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.146296] env[62460]: DEBUG oslo_vmware.api [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313985, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.242832] env[62460]: DEBUG nova.scheduler.client.report [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 983.352874] env[62460]: DEBUG nova.virt.hardware [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 983.353247] env[62460]: DEBUG nova.virt.hardware [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 983.353311] env[62460]: DEBUG nova.virt.hardware [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 983.353487] env[62460]: DEBUG nova.virt.hardware [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 983.353641] env[62460]: DEBUG nova.virt.hardware [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 983.353795] env[62460]: DEBUG nova.virt.hardware [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 983.354011] env[62460]: DEBUG nova.virt.hardware [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 983.354181] env[62460]: DEBUG nova.virt.hardware [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 983.354352] env[62460]: DEBUG nova.virt.hardware [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 983.354517] env[62460]: DEBUG nova.virt.hardware [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 983.354691] env[62460]: DEBUG nova.virt.hardware [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 983.359707] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b6bd8a70-b5fb-47c9-9f42-cbff43af776f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.375308] env[62460]: DEBUG oslo_vmware.api [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 983.375308] env[62460]: value = "task-1313987" [ 983.375308] env[62460]: _type = "Task" [ 983.375308] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.383274] env[62460]: DEBUG oslo_vmware.api [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313987, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.423626] env[62460]: DEBUG oslo_vmware.api [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Task: {'id': task-1313986, 'name': Rename_Task, 'duration_secs': 0.130874} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.423948] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 983.424206] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-89880b69-62d8-4df6-b0f1-7ac708700ca3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.429943] env[62460]: DEBUG oslo_vmware.api [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Waiting for the task: (returnval){ [ 983.429943] env[62460]: value = "task-1313988" [ 983.429943] env[62460]: _type = "Task" [ 983.429943] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.438062] env[62460]: DEBUG oslo_vmware.api [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Task: {'id': task-1313988, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.651502] env[62460]: DEBUG oslo_vmware.api [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313985, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.702746] env[62460]: DEBUG oslo_concurrency.lockutils [None req-56adbaa6-9771-4a23-bb49-025078b5223b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "interface-0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea-50d6a0c4-3681-4c1a-b42f-2987ee7a3bfb" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.703067] env[62460]: DEBUG oslo_concurrency.lockutils [None req-56adbaa6-9771-4a23-bb49-025078b5223b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "interface-0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea-50d6a0c4-3681-4c1a-b42f-2987ee7a3bfb" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.703456] env[62460]: DEBUG nova.objects.instance [None req-56adbaa6-9771-4a23-bb49-025078b5223b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lazy-loading 'flavor' on Instance uuid 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 983.748658] env[62460]: DEBUG oslo_concurrency.lockutils [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.225s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.749245] env[62460]: DEBUG nova.compute.manager [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 983.751991] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.434s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.753936] env[62460]: INFO nova.compute.claims [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 983.885587] env[62460]: DEBUG oslo_vmware.api [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313987, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.938894] env[62460]: DEBUG oslo_vmware.api [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Task: {'id': task-1313988, 'name': PowerOnVM_Task, 'duration_secs': 0.423649} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.940025] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 983.940025] env[62460]: INFO nova.compute.manager [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] Took 7.82 seconds to spawn the instance on the hypervisor. [ 983.940025] env[62460]: DEBUG nova.compute.manager [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 983.940636] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f4cb3b8-1ea9-449a-92aa-9c3d0ec0f123 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.149691] env[62460]: DEBUG oslo_vmware.api [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313985, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.258641] env[62460]: DEBUG nova.compute.utils [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 984.262283] env[62460]: DEBUG nova.compute.manager [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 984.262460] env[62460]: DEBUG nova.network.neutron [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 984.301515] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Acquiring lock "ddf1daf4-2772-48df-82c9-2192d012dedd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.301771] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lock "ddf1daf4-2772-48df-82c9-2192d012dedd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.304615] env[62460]: DEBUG nova.policy [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '632d7ca8cb4243f990d741a3f70f422a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '10ac54db1ef54e249a077bbb3dde4242', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 984.319162] env[62460]: DEBUG nova.objects.instance [None req-56adbaa6-9771-4a23-bb49-025078b5223b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lazy-loading 'pci_requests' on Instance uuid 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 984.386362] env[62460]: DEBUG oslo_vmware.api [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313987, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.460338] env[62460]: INFO nova.compute.manager [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] Took 12.70 seconds to build instance. [ 984.603308] env[62460]: DEBUG nova.network.neutron [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Successfully created port: 9dbf899d-26a8-45cd-994c-1913dd0b91d2 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 984.649022] env[62460]: DEBUG oslo_vmware.api [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313985, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.764236] env[62460]: DEBUG nova.compute.manager [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 984.803624] env[62460]: DEBUG nova.compute.manager [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 984.824025] env[62460]: DEBUG nova.objects.base [None req-56adbaa6-9771-4a23-bb49-025078b5223b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Object Instance<0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea> lazy-loaded attributes: flavor,pci_requests {{(pid=62460) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 984.824025] env[62460]: DEBUG nova.network.neutron [None req-56adbaa6-9771-4a23-bb49-025078b5223b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 984.892881] env[62460]: DEBUG oslo_vmware.api [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313987, 'name': ReconfigVM_Task, 'duration_secs': 1.249131} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.893254] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Updating instance '0d39a304-6a49-4d87-bfa2-4ba3b383578a' progress to 33 {{(pid=62460) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 984.900453] env[62460]: DEBUG nova.policy [None req-56adbaa6-9771-4a23-bb49-025078b5223b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd7b38fb0a53f4d078ae86d68e3eb6268', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '53fa06008e9f43488362895e7a143700', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 984.942536] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72b539bd-cdc7-4fdc-b029-3867adfda66e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.950567] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1daaae7f-a0ca-4b4c-8e9d-db102e4e5dd4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.981336] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b8a6861f-9e33-42c6-8730-9a3da5ea3981 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Lock "c772ecef-5cb5-4e69-a67c-0c487da4f1e3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.232s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.982342] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7628bb9c-35d8-4a33-aa2a-4268a3d9afb4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.989813] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc49737e-27b3-4dc9-aefc-f772964f96c1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.003574] env[62460]: DEBUG nova.compute.provider_tree [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 985.150435] env[62460]: DEBUG oslo_vmware.api [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1313985, 'name': PowerOnVM_Task, 'duration_secs': 2.392469} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.150749] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 985.151045] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a1c409ff-92e0-420e-acf2-b1293ff0f61a tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Updating instance '290500b5-d5b4-43b6-a843-20dacf561f4b' progress to 100 {{(pid=62460) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 985.322501] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.405136] env[62460]: DEBUG nova.virt.hardware [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 985.405471] env[62460]: DEBUG nova.virt.hardware [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 985.405563] env[62460]: DEBUG nova.virt.hardware [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 985.405751] env[62460]: DEBUG nova.virt.hardware [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 985.405905] env[62460]: DEBUG nova.virt.hardware [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 985.406076] env[62460]: DEBUG nova.virt.hardware [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 985.406297] env[62460]: DEBUG nova.virt.hardware [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 985.406467] env[62460]: DEBUG nova.virt.hardware [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 985.406642] env[62460]: DEBUG nova.virt.hardware [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 985.406813] env[62460]: DEBUG nova.virt.hardware [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 985.406992] env[62460]: DEBUG nova.virt.hardware [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 985.412270] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Reconfiguring VM instance instance-00000060 to detach disk 2000 {{(pid=62460) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 985.412556] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bb525336-d056-4de4-b87e-a057b372e50d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.431075] env[62460]: DEBUG oslo_vmware.api [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 985.431075] env[62460]: value = "task-1313989" [ 985.431075] env[62460]: _type = "Task" [ 985.431075] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.439488] env[62460]: DEBUG oslo_vmware.api [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313989, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.506936] env[62460]: DEBUG nova.scheduler.client.report [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 985.770292] env[62460]: DEBUG oslo_concurrency.lockutils [None req-596fc3e3-4e6f-4fa9-9198-11630e987896 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Acquiring lock "c772ecef-5cb5-4e69-a67c-0c487da4f1e3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.770579] env[62460]: DEBUG oslo_concurrency.lockutils [None req-596fc3e3-4e6f-4fa9-9198-11630e987896 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Lock "c772ecef-5cb5-4e69-a67c-0c487da4f1e3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.770806] env[62460]: DEBUG oslo_concurrency.lockutils [None req-596fc3e3-4e6f-4fa9-9198-11630e987896 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Acquiring lock "c772ecef-5cb5-4e69-a67c-0c487da4f1e3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.771027] env[62460]: DEBUG oslo_concurrency.lockutils [None req-596fc3e3-4e6f-4fa9-9198-11630e987896 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Lock "c772ecef-5cb5-4e69-a67c-0c487da4f1e3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.771204] env[62460]: DEBUG oslo_concurrency.lockutils [None req-596fc3e3-4e6f-4fa9-9198-11630e987896 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Lock "c772ecef-5cb5-4e69-a67c-0c487da4f1e3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.773329] env[62460]: INFO nova.compute.manager [None req-596fc3e3-4e6f-4fa9-9198-11630e987896 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] Terminating instance [ 985.775330] env[62460]: DEBUG nova.compute.manager [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 985.777896] env[62460]: DEBUG nova.compute.manager [None req-596fc3e3-4e6f-4fa9-9198-11630e987896 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 985.778159] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-596fc3e3-4e6f-4fa9-9198-11630e987896 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 985.779136] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b870584-8bd2-4c7e-812d-272f54d5ac47 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.786997] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-596fc3e3-4e6f-4fa9-9198-11630e987896 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 985.787249] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4a534614-9c24-43bf-a67f-a8dee7bcddf0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.793398] env[62460]: DEBUG oslo_vmware.api [None req-596fc3e3-4e6f-4fa9-9198-11630e987896 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Waiting for the task: (returnval){ [ 985.793398] env[62460]: value = "task-1313990" [ 985.793398] env[62460]: _type = "Task" [ 985.793398] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.800968] env[62460]: DEBUG nova.virt.hardware [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 985.801216] env[62460]: DEBUG nova.virt.hardware [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 985.801381] env[62460]: DEBUG nova.virt.hardware [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 985.801618] env[62460]: DEBUG nova.virt.hardware [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 985.801849] env[62460]: DEBUG nova.virt.hardware [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 985.802025] env[62460]: DEBUG nova.virt.hardware [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 985.802287] env[62460]: DEBUG nova.virt.hardware [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 985.802502] env[62460]: DEBUG nova.virt.hardware [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 985.802678] env[62460]: DEBUG nova.virt.hardware [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 985.802848] env[62460]: DEBUG nova.virt.hardware [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 985.803038] env[62460]: DEBUG nova.virt.hardware [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 985.803828] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7264dba2-b4b7-4158-8928-8000927d3ec7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.809347] env[62460]: DEBUG oslo_vmware.api [None req-596fc3e3-4e6f-4fa9-9198-11630e987896 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Task: {'id': task-1313990, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.814253] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1220c880-3381-4217-b83c-f0648cd55cad {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.940822] env[62460]: DEBUG oslo_vmware.api [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313989, 'name': ReconfigVM_Task, 'duration_secs': 0.159143} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.941166] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Reconfigured VM instance instance-00000060 to detach disk 2000 {{(pid=62460) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 985.941982] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a4709b9-e6da-49c2-8241-9e0f1ae6fffa {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.964400] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Reconfiguring VM instance instance-00000060 to attach disk [datastore1] 0d39a304-6a49-4d87-bfa2-4ba3b383578a/0d39a304-6a49-4d87-bfa2-4ba3b383578a.vmdk or device None with type thin {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 985.964731] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1d34596d-b981-446a-bde3-ae277c59d619 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.983563] env[62460]: DEBUG oslo_vmware.api [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 985.983563] env[62460]: value = "task-1313991" [ 985.983563] env[62460]: _type = "Task" [ 985.983563] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.991899] env[62460]: DEBUG oslo_vmware.api [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313991, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.011977] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.260s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.012617] env[62460]: DEBUG nova.compute.manager [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 986.015611] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.693s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.017155] env[62460]: INFO nova.compute.claims [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 986.130390] env[62460]: DEBUG nova.compute.manager [req-008a24b6-bb95-43a1-bc1c-8e93feb46926 req-4d910d18-df8d-4357-bb11-6e84754b9922 service nova] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Received event network-vif-plugged-9dbf899d-26a8-45cd-994c-1913dd0b91d2 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 986.130680] env[62460]: DEBUG oslo_concurrency.lockutils [req-008a24b6-bb95-43a1-bc1c-8e93feb46926 req-4d910d18-df8d-4357-bb11-6e84754b9922 service nova] Acquiring lock "4c5e60fe-6366-4fb9-975f-52345de67478-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.130987] env[62460]: DEBUG oslo_concurrency.lockutils [req-008a24b6-bb95-43a1-bc1c-8e93feb46926 req-4d910d18-df8d-4357-bb11-6e84754b9922 service nova] Lock "4c5e60fe-6366-4fb9-975f-52345de67478-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.131143] env[62460]: DEBUG oslo_concurrency.lockutils [req-008a24b6-bb95-43a1-bc1c-8e93feb46926 req-4d910d18-df8d-4357-bb11-6e84754b9922 service nova] Lock "4c5e60fe-6366-4fb9-975f-52345de67478-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.131326] env[62460]: DEBUG nova.compute.manager [req-008a24b6-bb95-43a1-bc1c-8e93feb46926 req-4d910d18-df8d-4357-bb11-6e84754b9922 service nova] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] No waiting events found dispatching network-vif-plugged-9dbf899d-26a8-45cd-994c-1913dd0b91d2 {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 986.131499] env[62460]: WARNING nova.compute.manager [req-008a24b6-bb95-43a1-bc1c-8e93feb46926 req-4d910d18-df8d-4357-bb11-6e84754b9922 service nova] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Received unexpected event network-vif-plugged-9dbf899d-26a8-45cd-994c-1913dd0b91d2 for instance with vm_state building and task_state spawning. [ 986.208560] env[62460]: DEBUG nova.network.neutron [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Successfully updated port: 9dbf899d-26a8-45cd-994c-1913dd0b91d2 {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 986.304824] env[62460]: DEBUG oslo_vmware.api [None req-596fc3e3-4e6f-4fa9-9198-11630e987896 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Task: {'id': task-1313990, 'name': PowerOffVM_Task, 'duration_secs': 0.188362} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.305176] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-596fc3e3-4e6f-4fa9-9198-11630e987896 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 986.305356] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-596fc3e3-4e6f-4fa9-9198-11630e987896 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 986.305618] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6f7f1fb3-9ee9-4a05-8527-7ff0af336220 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.323564] env[62460]: DEBUG nova.compute.manager [req-0efcde97-5663-4754-ba12-98f3c96bc2de req-070d618a-3487-4347-9d25-85875eb8dad5 service nova] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Received event network-vif-plugged-50d6a0c4-3681-4c1a-b42f-2987ee7a3bfb {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 986.323811] env[62460]: DEBUG oslo_concurrency.lockutils [req-0efcde97-5663-4754-ba12-98f3c96bc2de req-070d618a-3487-4347-9d25-85875eb8dad5 service nova] Acquiring lock "0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.324015] env[62460]: DEBUG oslo_concurrency.lockutils [req-0efcde97-5663-4754-ba12-98f3c96bc2de req-070d618a-3487-4347-9d25-85875eb8dad5 service nova] Lock "0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.324207] env[62460]: DEBUG oslo_concurrency.lockutils [req-0efcde97-5663-4754-ba12-98f3c96bc2de req-070d618a-3487-4347-9d25-85875eb8dad5 service nova] Lock "0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.324385] env[62460]: DEBUG nova.compute.manager [req-0efcde97-5663-4754-ba12-98f3c96bc2de req-070d618a-3487-4347-9d25-85875eb8dad5 service nova] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] No waiting events found dispatching network-vif-plugged-50d6a0c4-3681-4c1a-b42f-2987ee7a3bfb {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 986.324556] env[62460]: WARNING nova.compute.manager [req-0efcde97-5663-4754-ba12-98f3c96bc2de req-070d618a-3487-4347-9d25-85875eb8dad5 service nova] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Received unexpected event network-vif-plugged-50d6a0c4-3681-4c1a-b42f-2987ee7a3bfb for instance with vm_state active and task_state None. [ 986.367396] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-596fc3e3-4e6f-4fa9-9198-11630e987896 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 986.367651] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-596fc3e3-4e6f-4fa9-9198-11630e987896 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] Deleting contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 986.367858] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-596fc3e3-4e6f-4fa9-9198-11630e987896 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Deleting the datastore file [datastore2] c772ecef-5cb5-4e69-a67c-0c487da4f1e3 {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 986.368144] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-aacaefb4-2edb-4d41-9fff-c8c1e7decf20 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.374278] env[62460]: DEBUG oslo_vmware.api [None req-596fc3e3-4e6f-4fa9-9198-11630e987896 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Waiting for the task: (returnval){ [ 986.374278] env[62460]: value = "task-1313993" [ 986.374278] env[62460]: _type = "Task" [ 986.374278] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.381676] env[62460]: DEBUG oslo_vmware.api [None req-596fc3e3-4e6f-4fa9-9198-11630e987896 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Task: {'id': task-1313993, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.393594] env[62460]: DEBUG nova.network.neutron [None req-56adbaa6-9771-4a23-bb49-025078b5223b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Successfully updated port: 50d6a0c4-3681-4c1a-b42f-2987ee7a3bfb {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 986.493224] env[62460]: DEBUG oslo_vmware.api [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1313991, 'name': ReconfigVM_Task, 'duration_secs': 0.248343} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.493528] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Reconfigured VM instance instance-00000060 to attach disk [datastore1] 0d39a304-6a49-4d87-bfa2-4ba3b383578a/0d39a304-6a49-4d87-bfa2-4ba3b383578a.vmdk or device None with type thin {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 986.493813] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Updating instance '0d39a304-6a49-4d87-bfa2-4ba3b383578a' progress to 50 {{(pid=62460) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 986.522035] env[62460]: DEBUG nova.compute.utils [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 986.523220] env[62460]: DEBUG nova.compute.manager [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 986.523394] env[62460]: DEBUG nova.network.neutron [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 986.565425] env[62460]: DEBUG nova.policy [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a73816ab645e4cd49fabdbc4eeb31f97', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eee8bf23869e4680aec16e37f972175f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 986.710343] env[62460]: DEBUG oslo_concurrency.lockutils [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "refresh_cache-4c5e60fe-6366-4fb9-975f-52345de67478" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.710498] env[62460]: DEBUG oslo_concurrency.lockutils [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquired lock "refresh_cache-4c5e60fe-6366-4fb9-975f-52345de67478" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.710656] env[62460]: DEBUG nova.network.neutron [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 986.822638] env[62460]: DEBUG nova.network.neutron [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Successfully created port: 345c87ec-2bb1-420f-bc8c-845a795baad5 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 986.884640] env[62460]: DEBUG oslo_vmware.api [None req-596fc3e3-4e6f-4fa9-9198-11630e987896 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Task: {'id': task-1313993, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.136739} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.884923] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-596fc3e3-4e6f-4fa9-9198-11630e987896 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 986.885192] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-596fc3e3-4e6f-4fa9-9198-11630e987896 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] Deleted contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 986.885391] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-596fc3e3-4e6f-4fa9-9198-11630e987896 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 986.885723] env[62460]: INFO nova.compute.manager [None req-596fc3e3-4e6f-4fa9-9198-11630e987896 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] Took 1.11 seconds to destroy the instance on the hypervisor. [ 986.885831] env[62460]: DEBUG oslo.service.loopingcall [None req-596fc3e3-4e6f-4fa9-9198-11630e987896 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 986.886020] env[62460]: DEBUG nova.compute.manager [-] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 986.886289] env[62460]: DEBUG nova.network.neutron [-] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 986.896450] env[62460]: DEBUG oslo_concurrency.lockutils [None req-56adbaa6-9771-4a23-bb49-025078b5223b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "refresh_cache-0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.896567] env[62460]: DEBUG oslo_concurrency.lockutils [None req-56adbaa6-9771-4a23-bb49-025078b5223b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquired lock "refresh_cache-0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.896724] env[62460]: DEBUG nova.network.neutron [None req-56adbaa6-9771-4a23-bb49-025078b5223b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 987.000456] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9623267f-3afa-42f9-bb66-8c4c7a7a4be0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.021596] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e3e4208-080a-42df-9f0d-eddac4c740bb {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.039402] env[62460]: DEBUG nova.compute.manager [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 987.045909] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Updating instance '0d39a304-6a49-4d87-bfa2-4ba3b383578a' progress to 67 {{(pid=62460) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 987.191443] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed902bfa-0db3-4d1d-af61-42b7876e1ece {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.199790] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82c8cc4d-15f0-4ad1-8950-f908e60c6c6e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.233317] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-debca9bf-4593-4499-8ac4-6e34fa4bb716 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.241118] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50ecf99a-8528-4848-8823-38ba38c5a90a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.256386] env[62460]: DEBUG nova.compute.provider_tree [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 987.271739] env[62460]: DEBUG nova.network.neutron [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 987.276448] env[62460]: DEBUG oslo_concurrency.lockutils [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "290500b5-d5b4-43b6-a843-20dacf561f4b" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.276683] env[62460]: DEBUG oslo_concurrency.lockutils [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "290500b5-d5b4-43b6-a843-20dacf561f4b" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.276870] env[62460]: DEBUG nova.compute.manager [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Going to confirm migration 2 {{(pid=62460) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 987.437904] env[62460]: DEBUG nova.network.neutron [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Updating instance_info_cache with network_info: [{"id": "9dbf899d-26a8-45cd-994c-1913dd0b91d2", "address": "fa:16:3e:6b:25:55", "network": {"id": "c8b3d824-c8f0-4f7b-a91f-1987e04a5f09", "bridge": "br-int", "label": "tempest-ServersTestJSON-1769082812-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10ac54db1ef54e249a077bbb3dde4242", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "255460d5-71d4-4bfd-87f1-acc10085db7f", "external-id": "nsx-vlan-transportzone-152", "segmentation_id": 152, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9dbf899d-26", "ovs_interfaceid": "9dbf899d-26a8-45cd-994c-1913dd0b91d2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.443668] env[62460]: WARNING nova.network.neutron [None req-56adbaa6-9771-4a23-bb49-025078b5223b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] d3136e32-ad55-4b73-835c-8fa5f0480767 already exists in list: networks containing: ['d3136e32-ad55-4b73-835c-8fa5f0480767']. ignoring it [ 987.597362] env[62460]: DEBUG nova.network.neutron [-] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.616718] env[62460]: DEBUG nova.network.neutron [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Port 786ef177-843f-44d0-8920-2f4332e29154 binding to destination host cpu-1 is already ACTIVE {{(pid=62460) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 987.741106] env[62460]: DEBUG nova.network.neutron [None req-56adbaa6-9771-4a23-bb49-025078b5223b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Updating instance_info_cache with network_info: [{"id": "90a7d774-4d76-4525-8e6d-8c08fdb93fb4", "address": "fa:16:3e:5b:1c:b7", "network": {"id": "d3136e32-ad55-4b73-835c-8fa5f0480767", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1511414968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53fa06008e9f43488362895e7a143700", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "054fcd1e-638e-425a-a1de-78cb188ae026", "external-id": "nsx-vlan-transportzone-658", "segmentation_id": 658, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90a7d774-4d", "ovs_interfaceid": "90a7d774-4d76-4525-8e6d-8c08fdb93fb4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "50d6a0c4-3681-4c1a-b42f-2987ee7a3bfb", "address": "fa:16:3e:c6:6b:1c", "network": {"id": "d3136e32-ad55-4b73-835c-8fa5f0480767", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1511414968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53fa06008e9f43488362895e7a143700", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "054fcd1e-638e-425a-a1de-78cb188ae026", "external-id": "nsx-vlan-transportzone-658", "segmentation_id": 658, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap50d6a0c4-36", "ovs_interfaceid": "50d6a0c4-3681-4c1a-b42f-2987ee7a3bfb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.760016] env[62460]: DEBUG nova.scheduler.client.report [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 987.830825] env[62460]: DEBUG oslo_concurrency.lockutils [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "refresh_cache-290500b5-d5b4-43b6-a843-20dacf561f4b" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 987.831085] env[62460]: DEBUG oslo_concurrency.lockutils [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquired lock "refresh_cache-290500b5-d5b4-43b6-a843-20dacf561f4b" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.831311] env[62460]: DEBUG nova.network.neutron [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 987.831547] env[62460]: DEBUG nova.objects.instance [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lazy-loading 'info_cache' on Instance uuid 290500b5-d5b4-43b6-a843-20dacf561f4b {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 987.941397] env[62460]: DEBUG oslo_concurrency.lockutils [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Releasing lock "refresh_cache-4c5e60fe-6366-4fb9-975f-52345de67478" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.941397] env[62460]: DEBUG nova.compute.manager [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Instance network_info: |[{"id": "9dbf899d-26a8-45cd-994c-1913dd0b91d2", "address": "fa:16:3e:6b:25:55", "network": {"id": "c8b3d824-c8f0-4f7b-a91f-1987e04a5f09", "bridge": "br-int", "label": "tempest-ServersTestJSON-1769082812-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10ac54db1ef54e249a077bbb3dde4242", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "255460d5-71d4-4bfd-87f1-acc10085db7f", "external-id": "nsx-vlan-transportzone-152", "segmentation_id": 152, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9dbf899d-26", "ovs_interfaceid": "9dbf899d-26a8-45cd-994c-1913dd0b91d2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 987.941705] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6b:25:55', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '255460d5-71d4-4bfd-87f1-acc10085db7f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9dbf899d-26a8-45cd-994c-1913dd0b91d2', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 987.950207] env[62460]: DEBUG oslo.service.loopingcall [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 987.950428] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 987.950780] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-142dce95-0b5f-4e4f-9252-42af7af354f1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.971948] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 987.971948] env[62460]: value = "task-1313994" [ 987.971948] env[62460]: _type = "Task" [ 987.971948] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.979552] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313994, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.060255] env[62460]: DEBUG nova.compute.manager [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 988.086447] env[62460]: DEBUG nova.virt.hardware [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 988.086745] env[62460]: DEBUG nova.virt.hardware [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 988.086888] env[62460]: DEBUG nova.virt.hardware [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 988.087092] env[62460]: DEBUG nova.virt.hardware [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 988.087296] env[62460]: DEBUG nova.virt.hardware [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 988.087485] env[62460]: DEBUG nova.virt.hardware [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 988.087742] env[62460]: DEBUG nova.virt.hardware [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 988.087960] env[62460]: DEBUG nova.virt.hardware [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 988.088178] env[62460]: DEBUG nova.virt.hardware [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 988.088357] env[62460]: DEBUG nova.virt.hardware [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 988.088539] env[62460]: DEBUG nova.virt.hardware [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 988.089439] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8347e891-6acb-43cc-a575-61232d0141e5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.097197] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2d6a174-bcf5-4862-a369-2536d2cf575d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.101402] env[62460]: INFO nova.compute.manager [-] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] Took 1.22 seconds to deallocate network for instance. [ 988.247065] env[62460]: DEBUG oslo_concurrency.lockutils [None req-56adbaa6-9771-4a23-bb49-025078b5223b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Releasing lock "refresh_cache-0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.247065] env[62460]: DEBUG oslo_concurrency.lockutils [None req-56adbaa6-9771-4a23-bb49-025078b5223b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.247065] env[62460]: DEBUG oslo_concurrency.lockutils [None req-56adbaa6-9771-4a23-bb49-025078b5223b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquired lock "0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.247065] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0ff2c51-4cab-4a09-967f-b0615975ac69 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.265018] env[62460]: DEBUG nova.virt.hardware [None req-56adbaa6-9771-4a23-bb49-025078b5223b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 988.265018] env[62460]: DEBUG nova.virt.hardware [None req-56adbaa6-9771-4a23-bb49-025078b5223b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 988.265018] env[62460]: DEBUG nova.virt.hardware [None req-56adbaa6-9771-4a23-bb49-025078b5223b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 988.265018] env[62460]: DEBUG nova.virt.hardware [None req-56adbaa6-9771-4a23-bb49-025078b5223b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 988.265475] env[62460]: DEBUG nova.virt.hardware [None req-56adbaa6-9771-4a23-bb49-025078b5223b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 988.265789] env[62460]: DEBUG nova.virt.hardware [None req-56adbaa6-9771-4a23-bb49-025078b5223b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 988.266193] env[62460]: DEBUG nova.virt.hardware [None req-56adbaa6-9771-4a23-bb49-025078b5223b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 988.266526] env[62460]: DEBUG nova.virt.hardware [None req-56adbaa6-9771-4a23-bb49-025078b5223b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 988.269060] env[62460]: DEBUG nova.virt.hardware [None req-56adbaa6-9771-4a23-bb49-025078b5223b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 988.269060] env[62460]: DEBUG nova.virt.hardware [None req-56adbaa6-9771-4a23-bb49-025078b5223b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 988.269060] env[62460]: DEBUG nova.virt.hardware [None req-56adbaa6-9771-4a23-bb49-025078b5223b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 988.274474] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-56adbaa6-9771-4a23-bb49-025078b5223b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Reconfiguring VM to attach interface {{(pid=62460) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 988.275358] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.260s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.276103] env[62460]: DEBUG nova.compute.manager [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 988.280174] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3b4948aa-cb22-4455-bf43-477ea94abab6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.295264] env[62460]: DEBUG nova.network.neutron [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Successfully updated port: 345c87ec-2bb1-420f-bc8c-845a795baad5 {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 988.302828] env[62460]: DEBUG oslo_vmware.api [None req-56adbaa6-9771-4a23-bb49-025078b5223b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Waiting for the task: (returnval){ [ 988.302828] env[62460]: value = "task-1313995" [ 988.302828] env[62460]: _type = "Task" [ 988.302828] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.313179] env[62460]: DEBUG oslo_vmware.api [None req-56adbaa6-9771-4a23-bb49-025078b5223b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313995, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.481836] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313994, 'name': CreateVM_Task, 'duration_secs': 0.355178} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.481998] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 988.482703] env[62460]: DEBUG oslo_concurrency.lockutils [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.482889] env[62460]: DEBUG oslo_concurrency.lockutils [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.483254] env[62460]: DEBUG oslo_concurrency.lockutils [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 988.483523] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-455babcc-7407-499d-881e-7467e9cee9ca {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.487995] env[62460]: DEBUG oslo_vmware.api [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for the task: (returnval){ [ 988.487995] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]5255e6cb-573e-7545-9b8d-3ac5b5365dcb" [ 988.487995] env[62460]: _type = "Task" [ 988.487995] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.496290] env[62460]: DEBUG oslo_vmware.api [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5255e6cb-573e-7545-9b8d-3ac5b5365dcb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.616373] env[62460]: DEBUG oslo_concurrency.lockutils [None req-596fc3e3-4e6f-4fa9-9198-11630e987896 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.616738] env[62460]: DEBUG oslo_concurrency.lockutils [None req-596fc3e3-4e6f-4fa9-9198-11630e987896 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.616860] env[62460]: DEBUG nova.objects.instance [None req-596fc3e3-4e6f-4fa9-9198-11630e987896 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Lazy-loading 'resources' on Instance uuid c772ecef-5cb5-4e69-a67c-0c487da4f1e3 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 988.640233] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "0d39a304-6a49-4d87-bfa2-4ba3b383578a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.640496] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "0d39a304-6a49-4d87-bfa2-4ba3b383578a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.640695] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "0d39a304-6a49-4d87-bfa2-4ba3b383578a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.798203] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "refresh_cache-61f4666e-0588-4a00-a969-2349148caef2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.798463] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquired lock "refresh_cache-61f4666e-0588-4a00-a969-2349148caef2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.798660] env[62460]: DEBUG nova.network.neutron [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 988.801638] env[62460]: DEBUG nova.compute.utils [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 988.803860] env[62460]: DEBUG nova.compute.manager [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 988.804075] env[62460]: DEBUG nova.network.neutron [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 988.817063] env[62460]: DEBUG oslo_vmware.api [None req-56adbaa6-9771-4a23-bb49-025078b5223b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313995, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.849910] env[62460]: DEBUG nova.policy [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '45c39828856b472d894767f94f5d83ae', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cfc48e301fb04231b88ec28a062bda4e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 989.000266] env[62460]: DEBUG oslo_vmware.api [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5255e6cb-573e-7545-9b8d-3ac5b5365dcb, 'name': SearchDatastore_Task, 'duration_secs': 0.011055} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.000655] env[62460]: DEBUG oslo_concurrency.lockutils [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.000943] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 989.001268] env[62460]: DEBUG oslo_concurrency.lockutils [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.001449] env[62460]: DEBUG oslo_concurrency.lockutils [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.001685] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 989.002010] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a1504e67-06d9-4b25-a36b-4b7e53680e07 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.010563] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 989.010789] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 989.011589] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6d75e210-9e53-430a-aab3-713ee98e93ff {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.017572] env[62460]: DEBUG oslo_vmware.api [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for the task: (returnval){ [ 989.017572] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52afc9fa-14a2-5d97-1731-5188d1b305d3" [ 989.017572] env[62460]: _type = "Task" [ 989.017572] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.025075] env[62460]: DEBUG oslo_vmware.api [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52afc9fa-14a2-5d97-1731-5188d1b305d3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.109962] env[62460]: DEBUG nova.network.neutron [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Successfully created port: 5fe76c2c-4427-44ee-b05d-f0e6fbbd874e {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 989.252192] env[62460]: DEBUG nova.network.neutron [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Updating instance_info_cache with network_info: [{"id": "9f9c5b7b-922a-4387-b997-3f552dd85a50", "address": "fa:16:3e:78:80:d5", "network": {"id": "de789887-9cf9-4614-a7e1-dab430aaabaf", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-656400207-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0da056d93bdf40c39d6e82e457727ff6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5116f690-f825-4fee-8a47-42b073e716c5", "external-id": "nsx-vlan-transportzone-692", "segmentation_id": 692, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f9c5b7b-92", "ovs_interfaceid": "9f9c5b7b-922a-4387-b997-3f552dd85a50", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.304710] env[62460]: DEBUG nova.compute.manager [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 989.321340] env[62460]: DEBUG oslo_vmware.api [None req-56adbaa6-9771-4a23-bb49-025078b5223b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1313995, 'name': ReconfigVM_Task, 'duration_secs': 0.523876} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.321340] env[62460]: DEBUG oslo_concurrency.lockutils [None req-56adbaa6-9771-4a23-bb49-025078b5223b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Releasing lock "0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.321340] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-56adbaa6-9771-4a23-bb49-025078b5223b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Reconfigured VM to attach interface {{(pid=62460) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 989.346389] env[62460]: DEBUG nova.network.neutron [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 989.370898] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6eee41b1-84cd-4313-a591-a2b0a84e0152 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.378490] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b1ab29f-2f0d-414c-80b0-58bb66dd2ed7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.415304] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-545da7b6-9628-46b9-97fb-ea32158481c9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.422786] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a1b79a3-31d6-43d4-943c-c8fc3a9c24ad {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.435973] env[62460]: DEBUG nova.compute.provider_tree [None req-596fc3e3-4e6f-4fa9-9198-11630e987896 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 989.515568] env[62460]: DEBUG nova.network.neutron [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Updating instance_info_cache with network_info: [{"id": "345c87ec-2bb1-420f-bc8c-845a795baad5", "address": "fa:16:3e:43:36:b0", "network": {"id": "1dfc278d-e89e-4caf-9cfb-c302106cf974", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-868970047-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eee8bf23869e4680aec16e37f972175f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap345c87ec-2b", "ovs_interfaceid": "345c87ec-2bb1-420f-bc8c-845a795baad5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.527450] env[62460]: DEBUG oslo_vmware.api [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52afc9fa-14a2-5d97-1731-5188d1b305d3, 'name': SearchDatastore_Task, 'duration_secs': 0.008198} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.528735] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-579fe48b-f1f6-4a12-b6f6-900e5a64e690 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.534047] env[62460]: DEBUG oslo_vmware.api [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for the task: (returnval){ [ 989.534047] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52696d94-7109-92d3-c8c5-b5bd1733647e" [ 989.534047] env[62460]: _type = "Task" [ 989.534047] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.541221] env[62460]: DEBUG oslo_vmware.api [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52696d94-7109-92d3-c8c5-b5bd1733647e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.672911] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "refresh_cache-0d39a304-6a49-4d87-bfa2-4ba3b383578a" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.673196] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquired lock "refresh_cache-0d39a304-6a49-4d87-bfa2-4ba3b383578a" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.673310] env[62460]: DEBUG nova.network.neutron [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 989.754841] env[62460]: DEBUG oslo_concurrency.lockutils [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Releasing lock "refresh_cache-290500b5-d5b4-43b6-a843-20dacf561f4b" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.755136] env[62460]: DEBUG nova.objects.instance [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lazy-loading 'migration_context' on Instance uuid 290500b5-d5b4-43b6-a843-20dacf561f4b {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 989.826304] env[62460]: DEBUG oslo_concurrency.lockutils [None req-56adbaa6-9771-4a23-bb49-025078b5223b tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "interface-0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea-50d6a0c4-3681-4c1a-b42f-2987ee7a3bfb" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.123s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.939732] env[62460]: DEBUG nova.scheduler.client.report [None req-596fc3e3-4e6f-4fa9-9198-11630e987896 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 990.018160] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Releasing lock "refresh_cache-61f4666e-0588-4a00-a969-2349148caef2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.018619] env[62460]: DEBUG nova.compute.manager [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Instance network_info: |[{"id": "345c87ec-2bb1-420f-bc8c-845a795baad5", "address": "fa:16:3e:43:36:b0", "network": {"id": "1dfc278d-e89e-4caf-9cfb-c302106cf974", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-868970047-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eee8bf23869e4680aec16e37f972175f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap345c87ec-2b", "ovs_interfaceid": "345c87ec-2bb1-420f-bc8c-845a795baad5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 990.019221] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:43:36:b0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '418ddd3d-5f64-407e-8e0c-c8b81639bee9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '345c87ec-2bb1-420f-bc8c-845a795baad5', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 990.031799] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Creating folder: Project (eee8bf23869e4680aec16e37f972175f). Parent ref: group-v281134. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 990.032581] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4babd75f-cf31-473a-a441-59cfdbdb5b44 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.047952] env[62460]: DEBUG oslo_vmware.api [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52696d94-7109-92d3-c8c5-b5bd1733647e, 'name': SearchDatastore_Task, 'duration_secs': 0.00893} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.048356] env[62460]: DEBUG oslo_concurrency.lockutils [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.048724] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] 4c5e60fe-6366-4fb9-975f-52345de67478/4c5e60fe-6366-4fb9-975f-52345de67478.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 990.050279] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5828affe-5ba8-4bc0-bdcc-b26cb73af62a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.052796] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Created folder: Project (eee8bf23869e4680aec16e37f972175f) in parent group-v281134. [ 990.053090] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Creating folder: Instances. Parent ref: group-v281261. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 990.053389] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e485f158-1cff-4cc4-84d5-18b48ec06a5e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.060629] env[62460]: DEBUG oslo_vmware.api [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for the task: (returnval){ [ 990.060629] env[62460]: value = "task-1313997" [ 990.060629] env[62460]: _type = "Task" [ 990.060629] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.066265] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Created folder: Instances in parent group-v281261. [ 990.066600] env[62460]: DEBUG oslo.service.loopingcall [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 990.067245] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 990.067540] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-707eb8ff-ea07-4437-a5da-f0811500b652 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.097255] env[62460]: DEBUG oslo_vmware.api [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313997, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.101775] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 990.101775] env[62460]: value = "task-1313999" [ 990.101775] env[62460]: _type = "Task" [ 990.101775] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.108552] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313999, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.257620] env[62460]: DEBUG nova.objects.base [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Object Instance<290500b5-d5b4-43b6-a843-20dacf561f4b> lazy-loaded attributes: info_cache,migration_context {{(pid=62460) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 990.258635] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92b4d1c7-1f26-4aa5-a4fd-1f500e3c7784 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.281619] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a873ee8c-d86c-4c5a-ba81-3469fc3e6730 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.288357] env[62460]: DEBUG oslo_vmware.api [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 990.288357] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52a5e54e-b997-e63f-22af-49517cdab96b" [ 990.288357] env[62460]: _type = "Task" [ 990.288357] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.297017] env[62460]: DEBUG oslo_vmware.api [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52a5e54e-b997-e63f-22af-49517cdab96b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.316378] env[62460]: DEBUG nova.compute.manager [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 990.346801] env[62460]: DEBUG nova.virt.hardware [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 990.347133] env[62460]: DEBUG nova.virt.hardware [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 990.347341] env[62460]: DEBUG nova.virt.hardware [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 990.347550] env[62460]: DEBUG nova.virt.hardware [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 990.347683] env[62460]: DEBUG nova.virt.hardware [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 990.347875] env[62460]: DEBUG nova.virt.hardware [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 990.348121] env[62460]: DEBUG nova.virt.hardware [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 990.348298] env[62460]: DEBUG nova.virt.hardware [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 990.348478] env[62460]: DEBUG nova.virt.hardware [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 990.348653] env[62460]: DEBUG nova.virt.hardware [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 990.348840] env[62460]: DEBUG nova.virt.hardware [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 990.350250] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0607e712-7adb-4b5d-a279-81a626080be0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.361281] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-970059ce-db81-4083-9aa3-6d8a0a8eec69 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.429847] env[62460]: DEBUG nova.network.neutron [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Updating instance_info_cache with network_info: [{"id": "786ef177-843f-44d0-8920-2f4332e29154", "address": "fa:16:3e:09:59:ab", "network": {"id": "3c49720c-94c9-4e2f-bbb7-4266cc97faee", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1650281790-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfa5a65b0f614e769de5b3aa77bf869e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap786ef177-84", "ovs_interfaceid": "786ef177-843f-44d0-8920-2f4332e29154", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.446841] env[62460]: DEBUG oslo_concurrency.lockutils [None req-596fc3e3-4e6f-4fa9-9198-11630e987896 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.830s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.473747] env[62460]: INFO nova.scheduler.client.report [None req-596fc3e3-4e6f-4fa9-9198-11630e987896 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Deleted allocations for instance c772ecef-5cb5-4e69-a67c-0c487da4f1e3 [ 990.570738] env[62460]: DEBUG oslo_vmware.api [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1313997, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.48493} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.571099] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] 4c5e60fe-6366-4fb9-975f-52345de67478/4c5e60fe-6366-4fb9-975f-52345de67478.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 990.571331] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 990.571586] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-07ca85b4-83d3-40e3-8608-b9c933be3d07 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.577662] env[62460]: DEBUG oslo_vmware.api [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for the task: (returnval){ [ 990.577662] env[62460]: value = "task-1314000" [ 990.577662] env[62460]: _type = "Task" [ 990.577662] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.585144] env[62460]: DEBUG oslo_vmware.api [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1314000, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.610410] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1313999, 'name': CreateVM_Task, 'duration_secs': 0.354841} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.610599] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 990.611421] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.611481] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.611809] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 990.612082] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2fa170bd-043f-4d46-a8f9-d8f837db494b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.616526] env[62460]: DEBUG oslo_vmware.api [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 990.616526] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52859bb2-d71a-cd93-76f8-265c24b25b87" [ 990.616526] env[62460]: _type = "Task" [ 990.616526] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.617304] env[62460]: DEBUG nova.network.neutron [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Successfully updated port: 5fe76c2c-4427-44ee-b05d-f0e6fbbd874e {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 990.625937] env[62460]: DEBUG oslo_vmware.api [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52859bb2-d71a-cd93-76f8-265c24b25b87, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.798792] env[62460]: DEBUG oslo_vmware.api [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52a5e54e-b997-e63f-22af-49517cdab96b, 'name': SearchDatastore_Task, 'duration_secs': 0.008328} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.799123] env[62460]: DEBUG oslo_concurrency.lockutils [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.799364] env[62460]: DEBUG oslo_concurrency.lockutils [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.933130] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Releasing lock "refresh_cache-0d39a304-6a49-4d87-bfa2-4ba3b383578a" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.981064] env[62460]: DEBUG oslo_concurrency.lockutils [None req-596fc3e3-4e6f-4fa9-9198-11630e987896 tempest-ServerAddressesNegativeTestJSON-1736800550 tempest-ServerAddressesNegativeTestJSON-1736800550-project-member] Lock "c772ecef-5cb5-4e69-a67c-0c487da4f1e3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.210s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.086599] env[62460]: DEBUG oslo_vmware.api [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1314000, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.054944} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.086876] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 991.087637] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2096735e-3855-4a56-963a-0effa37e8008 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.108682] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] 4c5e60fe-6366-4fb9-975f-52345de67478/4c5e60fe-6366-4fb9-975f-52345de67478.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 991.108942] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e91611ae-dcb0-47e2-be49-ea67729590d2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.122982] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Acquiring lock "refresh_cache-ddf1daf4-2772-48df-82c9-2192d012dedd" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.123245] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Acquired lock "refresh_cache-ddf1daf4-2772-48df-82c9-2192d012dedd" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.123313] env[62460]: DEBUG nova.network.neutron [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 991.133113] env[62460]: DEBUG oslo_vmware.api [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52859bb2-d71a-cd93-76f8-265c24b25b87, 'name': SearchDatastore_Task, 'duration_secs': 0.076467} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.134841] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.135094] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 991.135337] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.135492] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.135678] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 991.135987] env[62460]: DEBUG oslo_vmware.api [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for the task: (returnval){ [ 991.135987] env[62460]: value = "task-1314001" [ 991.135987] env[62460]: _type = "Task" [ 991.135987] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.136415] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-611fe1a9-9035-4352-ae6c-8f7b4a57c1d3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.146400] env[62460]: DEBUG oslo_vmware.api [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1314001, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.151241] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 991.151370] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 991.152328] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bcbf6812-4de9-4ea2-8925-7e33f4ff3931 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.159265] env[62460]: DEBUG oslo_vmware.api [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 991.159265] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]525ab871-202d-9954-3f49-a8827c13e834" [ 991.159265] env[62460]: _type = "Task" [ 991.159265] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.167094] env[62460]: DEBUG oslo_vmware.api [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]525ab871-202d-9954-3f49-a8827c13e834, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.297119] env[62460]: DEBUG nova.compute.manager [req-387fb1c4-7d93-43af-8997-8e57369c3167 req-990d598a-0f6c-4fc5-b634-e886a77875fa service nova] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Received event network-changed-9dbf899d-26a8-45cd-994c-1913dd0b91d2 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 991.297352] env[62460]: DEBUG nova.compute.manager [req-387fb1c4-7d93-43af-8997-8e57369c3167 req-990d598a-0f6c-4fc5-b634-e886a77875fa service nova] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Refreshing instance network info cache due to event network-changed-9dbf899d-26a8-45cd-994c-1913dd0b91d2. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 991.297595] env[62460]: DEBUG oslo_concurrency.lockutils [req-387fb1c4-7d93-43af-8997-8e57369c3167 req-990d598a-0f6c-4fc5-b634-e886a77875fa service nova] Acquiring lock "refresh_cache-4c5e60fe-6366-4fb9-975f-52345de67478" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.297758] env[62460]: DEBUG oslo_concurrency.lockutils [req-387fb1c4-7d93-43af-8997-8e57369c3167 req-990d598a-0f6c-4fc5-b634-e886a77875fa service nova] Acquired lock "refresh_cache-4c5e60fe-6366-4fb9-975f-52345de67478" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.298015] env[62460]: DEBUG nova.network.neutron [req-387fb1c4-7d93-43af-8997-8e57369c3167 req-990d598a-0f6c-4fc5-b634-e886a77875fa service nova] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Refreshing network info cache for port 9dbf899d-26a8-45cd-994c-1913dd0b91d2 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 991.320096] env[62460]: DEBUG nova.compute.manager [req-c521a0b3-722e-4744-a897-2ec3bd47a881 req-1335097c-85ef-4cf6-80fb-272ded814435 service nova] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Received event network-changed-50d6a0c4-3681-4c1a-b42f-2987ee7a3bfb {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 991.320348] env[62460]: DEBUG nova.compute.manager [req-c521a0b3-722e-4744-a897-2ec3bd47a881 req-1335097c-85ef-4cf6-80fb-272ded814435 service nova] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Refreshing instance network info cache due to event network-changed-50d6a0c4-3681-4c1a-b42f-2987ee7a3bfb. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 991.320523] env[62460]: DEBUG oslo_concurrency.lockutils [req-c521a0b3-722e-4744-a897-2ec3bd47a881 req-1335097c-85ef-4cf6-80fb-272ded814435 service nova] Acquiring lock "refresh_cache-0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.320677] env[62460]: DEBUG oslo_concurrency.lockutils [req-c521a0b3-722e-4744-a897-2ec3bd47a881 req-1335097c-85ef-4cf6-80fb-272ded814435 service nova] Acquired lock "refresh_cache-0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.320844] env[62460]: DEBUG nova.network.neutron [req-c521a0b3-722e-4744-a897-2ec3bd47a881 req-1335097c-85ef-4cf6-80fb-272ded814435 service nova] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Refreshing network info cache for port 50d6a0c4-3681-4c1a-b42f-2987ee7a3bfb {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 991.444358] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab4f15ca-f3c4-44a7-95ef-23dd70341ed7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.454482] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb9035e0-c97b-4563-84d7-6ea03ac81ba1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.460953] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b806ab3c-911c-4ca3-8250-0d6c44e5e746 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.501575] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2adf9104-dfb5-4c2e-964f-62ff8458a2d1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.505330] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30d5b691-4059-4825-8b6c-3581e39dc111 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.514481] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8b2a312-feec-45c4-9e40-ca6df4cf3448 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.518410] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Updating instance '0d39a304-6a49-4d87-bfa2-4ba3b383578a' progress to 83 {{(pid=62460) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 991.531201] env[62460]: DEBUG nova.compute.provider_tree [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 991.650702] env[62460]: DEBUG oslo_vmware.api [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1314001, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.668037] env[62460]: DEBUG oslo_vmware.api [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]525ab871-202d-9954-3f49-a8827c13e834, 'name': SearchDatastore_Task, 'duration_secs': 0.015493} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.668805] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-92ca35cc-dcb1-47fb-8f09-4ec50c23c23d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.671512] env[62460]: DEBUG nova.network.neutron [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 991.676226] env[62460]: DEBUG oslo_vmware.api [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 991.676226] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52f3f790-4da4-95b7-97a0-f1fa70b561d0" [ 991.676226] env[62460]: _type = "Task" [ 991.676226] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.683243] env[62460]: DEBUG oslo_vmware.api [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52f3f790-4da4-95b7-97a0-f1fa70b561d0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.926604] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9747b473-768f-4dbc-884d-4e0e4644fdc1 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "interface-0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea-50d6a0c4-3681-4c1a-b42f-2987ee7a3bfb" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.926876] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9747b473-768f-4dbc-884d-4e0e4644fdc1 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "interface-0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea-50d6a0c4-3681-4c1a-b42f-2987ee7a3bfb" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.024935] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 992.025295] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-080b8787-caa3-4d6d-9792-55424ee06331 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.032886] env[62460]: DEBUG oslo_vmware.api [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 992.032886] env[62460]: value = "task-1314002" [ 992.032886] env[62460]: _type = "Task" [ 992.032886] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.033660] env[62460]: DEBUG nova.scheduler.client.report [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 992.045337] env[62460]: DEBUG oslo_vmware.api [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1314002, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.116582] env[62460]: DEBUG nova.network.neutron [req-387fb1c4-7d93-43af-8997-8e57369c3167 req-990d598a-0f6c-4fc5-b634-e886a77875fa service nova] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Updated VIF entry in instance network info cache for port 9dbf899d-26a8-45cd-994c-1913dd0b91d2. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 992.116980] env[62460]: DEBUG nova.network.neutron [req-387fb1c4-7d93-43af-8997-8e57369c3167 req-990d598a-0f6c-4fc5-b634-e886a77875fa service nova] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Updating instance_info_cache with network_info: [{"id": "9dbf899d-26a8-45cd-994c-1913dd0b91d2", "address": "fa:16:3e:6b:25:55", "network": {"id": "c8b3d824-c8f0-4f7b-a91f-1987e04a5f09", "bridge": "br-int", "label": "tempest-ServersTestJSON-1769082812-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10ac54db1ef54e249a077bbb3dde4242", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "255460d5-71d4-4bfd-87f1-acc10085db7f", "external-id": "nsx-vlan-transportzone-152", "segmentation_id": 152, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9dbf899d-26", "ovs_interfaceid": "9dbf899d-26a8-45cd-994c-1913dd0b91d2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.123871] env[62460]: DEBUG nova.network.neutron [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Updating instance_info_cache with network_info: [{"id": "5fe76c2c-4427-44ee-b05d-f0e6fbbd874e", "address": "fa:16:3e:d6:e3:16", "network": {"id": "4cb03486-3bea-4c6a-a019-37486f271107", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-2094026403-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfc48e301fb04231b88ec28a062bda4e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c7d2575f-b92f-44ec-a863-634cb76631a2", "external-id": "nsx-vlan-transportzone-794", "segmentation_id": 794, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5fe76c2c-44", "ovs_interfaceid": "5fe76c2c-4427-44ee-b05d-f0e6fbbd874e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.151489] env[62460]: DEBUG oslo_vmware.api [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1314001, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.176719] env[62460]: DEBUG nova.network.neutron [req-c521a0b3-722e-4744-a897-2ec3bd47a881 req-1335097c-85ef-4cf6-80fb-272ded814435 service nova] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Updated VIF entry in instance network info cache for port 50d6a0c4-3681-4c1a-b42f-2987ee7a3bfb. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 992.177207] env[62460]: DEBUG nova.network.neutron [req-c521a0b3-722e-4744-a897-2ec3bd47a881 req-1335097c-85ef-4cf6-80fb-272ded814435 service nova] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Updating instance_info_cache with network_info: [{"id": "90a7d774-4d76-4525-8e6d-8c08fdb93fb4", "address": "fa:16:3e:5b:1c:b7", "network": {"id": "d3136e32-ad55-4b73-835c-8fa5f0480767", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1511414968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53fa06008e9f43488362895e7a143700", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "054fcd1e-638e-425a-a1de-78cb188ae026", "external-id": "nsx-vlan-transportzone-658", "segmentation_id": 658, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90a7d774-4d", "ovs_interfaceid": "90a7d774-4d76-4525-8e6d-8c08fdb93fb4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "50d6a0c4-3681-4c1a-b42f-2987ee7a3bfb", "address": "fa:16:3e:c6:6b:1c", "network": {"id": "d3136e32-ad55-4b73-835c-8fa5f0480767", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1511414968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53fa06008e9f43488362895e7a143700", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "054fcd1e-638e-425a-a1de-78cb188ae026", "external-id": "nsx-vlan-transportzone-658", "segmentation_id": 658, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap50d6a0c4-36", "ovs_interfaceid": "50d6a0c4-3681-4c1a-b42f-2987ee7a3bfb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.189255] env[62460]: DEBUG oslo_vmware.api [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52f3f790-4da4-95b7-97a0-f1fa70b561d0, 'name': SearchDatastore_Task, 'duration_secs': 0.010889} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.189361] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.190207] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 61f4666e-0588-4a00-a969-2349148caef2/61f4666e-0588-4a00-a969-2349148caef2.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 992.190207] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d94301ea-2471-42d7-b1f9-764d1dc3caa7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.196477] env[62460]: DEBUG oslo_vmware.api [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 992.196477] env[62460]: value = "task-1314003" [ 992.196477] env[62460]: _type = "Task" [ 992.196477] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.205232] env[62460]: DEBUG oslo_vmware.api [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314003, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.430757] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9747b473-768f-4dbc-884d-4e0e4644fdc1 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 992.431096] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9747b473-768f-4dbc-884d-4e0e4644fdc1 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquired lock "0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.432106] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ce21cc0-84e0-4961-8aa0-041f0c61ec3e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.450042] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18b3501e-0d72-4f3a-91fd-1dc33af9b549 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.477416] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-9747b473-768f-4dbc-884d-4e0e4644fdc1 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Reconfiguring VM to detach interface {{(pid=62460) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 992.477785] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c7c6d140-ca02-4a1a-a75d-c5dc7c2bf985 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.497104] env[62460]: DEBUG oslo_vmware.api [None req-9747b473-768f-4dbc-884d-4e0e4644fdc1 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Waiting for the task: (returnval){ [ 992.497104] env[62460]: value = "task-1314004" [ 992.497104] env[62460]: _type = "Task" [ 992.497104] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.508819] env[62460]: DEBUG oslo_vmware.api [None req-9747b473-768f-4dbc-884d-4e0e4644fdc1 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1314004, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.551102] env[62460]: DEBUG oslo_vmware.api [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1314002, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.619763] env[62460]: DEBUG oslo_concurrency.lockutils [req-387fb1c4-7d93-43af-8997-8e57369c3167 req-990d598a-0f6c-4fc5-b634-e886a77875fa service nova] Releasing lock "refresh_cache-4c5e60fe-6366-4fb9-975f-52345de67478" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.620168] env[62460]: DEBUG nova.compute.manager [req-387fb1c4-7d93-43af-8997-8e57369c3167 req-990d598a-0f6c-4fc5-b634-e886a77875fa service nova] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] Received event network-vif-deleted-460727fc-e34d-424f-9770-d8dbea3ed624 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 992.626019] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Releasing lock "refresh_cache-ddf1daf4-2772-48df-82c9-2192d012dedd" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.626360] env[62460]: DEBUG nova.compute.manager [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Instance network_info: |[{"id": "5fe76c2c-4427-44ee-b05d-f0e6fbbd874e", "address": "fa:16:3e:d6:e3:16", "network": {"id": "4cb03486-3bea-4c6a-a019-37486f271107", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-2094026403-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfc48e301fb04231b88ec28a062bda4e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c7d2575f-b92f-44ec-a863-634cb76631a2", "external-id": "nsx-vlan-transportzone-794", "segmentation_id": 794, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5fe76c2c-44", "ovs_interfaceid": "5fe76c2c-4427-44ee-b05d-f0e6fbbd874e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 992.626854] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d6:e3:16', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c7d2575f-b92f-44ec-a863-634cb76631a2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5fe76c2c-4427-44ee-b05d-f0e6fbbd874e', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 992.634510] env[62460]: DEBUG oslo.service.loopingcall [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 992.635542] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 992.635542] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-00253bf6-875a-40c9-aad0-51b107f67bae {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.661499] env[62460]: DEBUG oslo_vmware.api [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1314001, 'name': ReconfigVM_Task, 'duration_secs': 1.317965} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.662706] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Reconfigured VM instance instance-00000062 to attach disk [datastore2] 4c5e60fe-6366-4fb9-975f-52345de67478/4c5e60fe-6366-4fb9-975f-52345de67478.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 992.663376] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 992.663376] env[62460]: value = "task-1314005" [ 992.663376] env[62460]: _type = "Task" [ 992.663376] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.663587] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8146dff1-5195-4eba-bc40-c204acd647a5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.673821] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1314005, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.675336] env[62460]: DEBUG oslo_vmware.api [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for the task: (returnval){ [ 992.675336] env[62460]: value = "task-1314006" [ 992.675336] env[62460]: _type = "Task" [ 992.675336] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.683560] env[62460]: DEBUG oslo_concurrency.lockutils [req-c521a0b3-722e-4744-a897-2ec3bd47a881 req-1335097c-85ef-4cf6-80fb-272ded814435 service nova] Releasing lock "refresh_cache-0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.683857] env[62460]: DEBUG nova.compute.manager [req-c521a0b3-722e-4744-a897-2ec3bd47a881 req-1335097c-85ef-4cf6-80fb-272ded814435 service nova] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Received event network-vif-plugged-345c87ec-2bb1-420f-bc8c-845a795baad5 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 992.684064] env[62460]: DEBUG oslo_concurrency.lockutils [req-c521a0b3-722e-4744-a897-2ec3bd47a881 req-1335097c-85ef-4cf6-80fb-272ded814435 service nova] Acquiring lock "61f4666e-0588-4a00-a969-2349148caef2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.684290] env[62460]: DEBUG oslo_concurrency.lockutils [req-c521a0b3-722e-4744-a897-2ec3bd47a881 req-1335097c-85ef-4cf6-80fb-272ded814435 service nova] Lock "61f4666e-0588-4a00-a969-2349148caef2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.684459] env[62460]: DEBUG oslo_concurrency.lockutils [req-c521a0b3-722e-4744-a897-2ec3bd47a881 req-1335097c-85ef-4cf6-80fb-272ded814435 service nova] Lock "61f4666e-0588-4a00-a969-2349148caef2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.684635] env[62460]: DEBUG nova.compute.manager [req-c521a0b3-722e-4744-a897-2ec3bd47a881 req-1335097c-85ef-4cf6-80fb-272ded814435 service nova] [instance: 61f4666e-0588-4a00-a969-2349148caef2] No waiting events found dispatching network-vif-plugged-345c87ec-2bb1-420f-bc8c-845a795baad5 {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 992.684811] env[62460]: WARNING nova.compute.manager [req-c521a0b3-722e-4744-a897-2ec3bd47a881 req-1335097c-85ef-4cf6-80fb-272ded814435 service nova] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Received unexpected event network-vif-plugged-345c87ec-2bb1-420f-bc8c-845a795baad5 for instance with vm_state building and task_state spawning. [ 992.684987] env[62460]: DEBUG nova.compute.manager [req-c521a0b3-722e-4744-a897-2ec3bd47a881 req-1335097c-85ef-4cf6-80fb-272ded814435 service nova] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Received event network-changed-345c87ec-2bb1-420f-bc8c-845a795baad5 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 992.685171] env[62460]: DEBUG nova.compute.manager [req-c521a0b3-722e-4744-a897-2ec3bd47a881 req-1335097c-85ef-4cf6-80fb-272ded814435 service nova] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Refreshing instance network info cache due to event network-changed-345c87ec-2bb1-420f-bc8c-845a795baad5. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 992.685364] env[62460]: DEBUG oslo_concurrency.lockutils [req-c521a0b3-722e-4744-a897-2ec3bd47a881 req-1335097c-85ef-4cf6-80fb-272ded814435 service nova] Acquiring lock "refresh_cache-61f4666e-0588-4a00-a969-2349148caef2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 992.685507] env[62460]: DEBUG oslo_concurrency.lockutils [req-c521a0b3-722e-4744-a897-2ec3bd47a881 req-1335097c-85ef-4cf6-80fb-272ded814435 service nova] Acquired lock "refresh_cache-61f4666e-0588-4a00-a969-2349148caef2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.685668] env[62460]: DEBUG nova.network.neutron [req-c521a0b3-722e-4744-a897-2ec3bd47a881 req-1335097c-85ef-4cf6-80fb-272ded814435 service nova] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Refreshing network info cache for port 345c87ec-2bb1-420f-bc8c-845a795baad5 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 992.686932] env[62460]: DEBUG oslo_vmware.api [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1314006, 'name': Rename_Task} progress is 6%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.706902] env[62460]: DEBUG oslo_vmware.api [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314003, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.009090] env[62460]: DEBUG oslo_vmware.api [None req-9747b473-768f-4dbc-884d-4e0e4644fdc1 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1314004, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.049655] env[62460]: DEBUG oslo_vmware.api [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1314002, 'name': PowerOnVM_Task, 'duration_secs': 0.875999} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.050037] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 993.050340] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a5577909-8ddc-4cd4-b21b-308df77972d5 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Updating instance '0d39a304-6a49-4d87-bfa2-4ba3b383578a' progress to 100 {{(pid=62460) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 993.055602] env[62460]: DEBUG oslo_concurrency.lockutils [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.256s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.175802] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1314005, 'name': CreateVM_Task} progress is 99%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.185316] env[62460]: DEBUG oslo_vmware.api [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1314006, 'name': Rename_Task, 'duration_secs': 0.159375} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.185316] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 993.185316] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-193a2be0-b1c3-458c-83c0-d9543cf128dd {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.191150] env[62460]: DEBUG oslo_vmware.api [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for the task: (returnval){ [ 993.191150] env[62460]: value = "task-1314007" [ 993.191150] env[62460]: _type = "Task" [ 993.191150] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.198615] env[62460]: DEBUG oslo_vmware.api [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1314007, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.208176] env[62460]: DEBUG oslo_vmware.api [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314003, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.538351} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.208176] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 61f4666e-0588-4a00-a969-2349148caef2/61f4666e-0588-4a00-a969-2349148caef2.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 993.208176] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 993.208176] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9870bc67-5760-4de8-b274-a84e80d88ae3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.214600] env[62460]: DEBUG oslo_vmware.api [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 993.214600] env[62460]: value = "task-1314008" [ 993.214600] env[62460]: _type = "Task" [ 993.214600] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.225516] env[62460]: DEBUG oslo_vmware.api [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314008, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.351892] env[62460]: DEBUG nova.compute.manager [req-26873b8c-4aea-435f-8236-117102c6ba21 req-6da18bae-1d66-4e1b-93e1-a48e3d7cba91 service nova] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Received event network-vif-plugged-5fe76c2c-4427-44ee-b05d-f0e6fbbd874e {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 993.353621] env[62460]: DEBUG oslo_concurrency.lockutils [req-26873b8c-4aea-435f-8236-117102c6ba21 req-6da18bae-1d66-4e1b-93e1-a48e3d7cba91 service nova] Acquiring lock "ddf1daf4-2772-48df-82c9-2192d012dedd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.353621] env[62460]: DEBUG oslo_concurrency.lockutils [req-26873b8c-4aea-435f-8236-117102c6ba21 req-6da18bae-1d66-4e1b-93e1-a48e3d7cba91 service nova] Lock "ddf1daf4-2772-48df-82c9-2192d012dedd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.353621] env[62460]: DEBUG oslo_concurrency.lockutils [req-26873b8c-4aea-435f-8236-117102c6ba21 req-6da18bae-1d66-4e1b-93e1-a48e3d7cba91 service nova] Lock "ddf1daf4-2772-48df-82c9-2192d012dedd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.353621] env[62460]: DEBUG nova.compute.manager [req-26873b8c-4aea-435f-8236-117102c6ba21 req-6da18bae-1d66-4e1b-93e1-a48e3d7cba91 service nova] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] No waiting events found dispatching network-vif-plugged-5fe76c2c-4427-44ee-b05d-f0e6fbbd874e {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 993.353621] env[62460]: WARNING nova.compute.manager [req-26873b8c-4aea-435f-8236-117102c6ba21 req-6da18bae-1d66-4e1b-93e1-a48e3d7cba91 service nova] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Received unexpected event network-vif-plugged-5fe76c2c-4427-44ee-b05d-f0e6fbbd874e for instance with vm_state building and task_state spawning. [ 993.353621] env[62460]: DEBUG nova.compute.manager [req-26873b8c-4aea-435f-8236-117102c6ba21 req-6da18bae-1d66-4e1b-93e1-a48e3d7cba91 service nova] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Received event network-changed-5fe76c2c-4427-44ee-b05d-f0e6fbbd874e {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 993.353621] env[62460]: DEBUG nova.compute.manager [req-26873b8c-4aea-435f-8236-117102c6ba21 req-6da18bae-1d66-4e1b-93e1-a48e3d7cba91 service nova] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Refreshing instance network info cache due to event network-changed-5fe76c2c-4427-44ee-b05d-f0e6fbbd874e. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 993.353621] env[62460]: DEBUG oslo_concurrency.lockutils [req-26873b8c-4aea-435f-8236-117102c6ba21 req-6da18bae-1d66-4e1b-93e1-a48e3d7cba91 service nova] Acquiring lock "refresh_cache-ddf1daf4-2772-48df-82c9-2192d012dedd" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.354050] env[62460]: DEBUG oslo_concurrency.lockutils [req-26873b8c-4aea-435f-8236-117102c6ba21 req-6da18bae-1d66-4e1b-93e1-a48e3d7cba91 service nova] Acquired lock "refresh_cache-ddf1daf4-2772-48df-82c9-2192d012dedd" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.354050] env[62460]: DEBUG nova.network.neutron [req-26873b8c-4aea-435f-8236-117102c6ba21 req-6da18bae-1d66-4e1b-93e1-a48e3d7cba91 service nova] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Refreshing network info cache for port 5fe76c2c-4427-44ee-b05d-f0e6fbbd874e {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 993.436440] env[62460]: DEBUG nova.network.neutron [req-c521a0b3-722e-4744-a897-2ec3bd47a881 req-1335097c-85ef-4cf6-80fb-272ded814435 service nova] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Updated VIF entry in instance network info cache for port 345c87ec-2bb1-420f-bc8c-845a795baad5. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 993.436889] env[62460]: DEBUG nova.network.neutron [req-c521a0b3-722e-4744-a897-2ec3bd47a881 req-1335097c-85ef-4cf6-80fb-272ded814435 service nova] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Updating instance_info_cache with network_info: [{"id": "345c87ec-2bb1-420f-bc8c-845a795baad5", "address": "fa:16:3e:43:36:b0", "network": {"id": "1dfc278d-e89e-4caf-9cfb-c302106cf974", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-868970047-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eee8bf23869e4680aec16e37f972175f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap345c87ec-2b", "ovs_interfaceid": "345c87ec-2bb1-420f-bc8c-845a795baad5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.507438] env[62460]: DEBUG oslo_vmware.api [None req-9747b473-768f-4dbc-884d-4e0e4644fdc1 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1314004, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.616623] env[62460]: INFO nova.scheduler.client.report [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Deleted allocation for migration 6c432a03-9e05-49cc-bf18-7f0bc8bde6eb [ 993.675664] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1314005, 'name': CreateVM_Task, 'duration_secs': 0.525293} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.675835] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 993.676514] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.676693] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.677036] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 993.677327] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ba68a5af-aef6-4e62-9fbe-1548e7204d4c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.682219] env[62460]: DEBUG oslo_vmware.api [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Waiting for the task: (returnval){ [ 993.682219] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52e9350e-58e6-7f58-e9e3-2ad0d695e547" [ 993.682219] env[62460]: _type = "Task" [ 993.682219] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.689397] env[62460]: DEBUG oslo_vmware.api [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52e9350e-58e6-7f58-e9e3-2ad0d695e547, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.698947] env[62460]: DEBUG oslo_vmware.api [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1314007, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.722922] env[62460]: DEBUG oslo_vmware.api [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314008, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063466} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.723200] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 993.723989] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f619b50-cd15-4620-ad16-e54fa9f9e0c4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.745951] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] 61f4666e-0588-4a00-a969-2349148caef2/61f4666e-0588-4a00-a969-2349148caef2.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 993.746200] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d1788d54-ff37-4956-88ba-c5c04c98bf3e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.765250] env[62460]: DEBUG oslo_vmware.api [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 993.765250] env[62460]: value = "task-1314009" [ 993.765250] env[62460]: _type = "Task" [ 993.765250] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.772657] env[62460]: DEBUG oslo_vmware.api [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314009, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.944037] env[62460]: DEBUG oslo_concurrency.lockutils [req-c521a0b3-722e-4744-a897-2ec3bd47a881 req-1335097c-85ef-4cf6-80fb-272ded814435 service nova] Releasing lock "refresh_cache-61f4666e-0588-4a00-a969-2349148caef2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.008912] env[62460]: DEBUG oslo_vmware.api [None req-9747b473-768f-4dbc-884d-4e0e4644fdc1 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1314004, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.078851] env[62460]: DEBUG nova.network.neutron [req-26873b8c-4aea-435f-8236-117102c6ba21 req-6da18bae-1d66-4e1b-93e1-a48e3d7cba91 service nova] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Updated VIF entry in instance network info cache for port 5fe76c2c-4427-44ee-b05d-f0e6fbbd874e. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 994.079249] env[62460]: DEBUG nova.network.neutron [req-26873b8c-4aea-435f-8236-117102c6ba21 req-6da18bae-1d66-4e1b-93e1-a48e3d7cba91 service nova] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Updating instance_info_cache with network_info: [{"id": "5fe76c2c-4427-44ee-b05d-f0e6fbbd874e", "address": "fa:16:3e:d6:e3:16", "network": {"id": "4cb03486-3bea-4c6a-a019-37486f271107", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-2094026403-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfc48e301fb04231b88ec28a062bda4e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c7d2575f-b92f-44ec-a863-634cb76631a2", "external-id": "nsx-vlan-transportzone-794", "segmentation_id": 794, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5fe76c2c-44", "ovs_interfaceid": "5fe76c2c-4427-44ee-b05d-f0e6fbbd874e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 994.122548] env[62460]: DEBUG oslo_concurrency.lockutils [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "290500b5-d5b4-43b6-a843-20dacf561f4b" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.846s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.195457] env[62460]: DEBUG oslo_vmware.api [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52e9350e-58e6-7f58-e9e3-2ad0d695e547, 'name': SearchDatastore_Task, 'duration_secs': 0.042248} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.198696] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.198952] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 994.199221] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 994.199374] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.199559] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 994.200235] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ff6890c0-a0bd-4849-8bf9-8e9e72e9fca3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.207472] env[62460]: DEBUG oslo_vmware.api [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1314007, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.216562] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 994.216748] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 994.217516] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-77ce4ae9-56a1-424a-b506-2ffc676f189e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.222667] env[62460]: DEBUG oslo_vmware.api [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Waiting for the task: (returnval){ [ 994.222667] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]5285f332-fa0d-1dd7-e1f2-a7aaa8fda2b4" [ 994.222667] env[62460]: _type = "Task" [ 994.222667] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.230420] env[62460]: DEBUG oslo_vmware.api [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5285f332-fa0d-1dd7-e1f2-a7aaa8fda2b4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.275138] env[62460]: DEBUG oslo_vmware.api [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314009, 'name': ReconfigVM_Task, 'duration_secs': 0.255767} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.275428] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Reconfigured VM instance instance-00000063 to attach disk [datastore1] 61f4666e-0588-4a00-a969-2349148caef2/61f4666e-0588-4a00-a969-2349148caef2.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 994.276060] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5990a88d-f9e3-42c0-a692-8d415c3dd314 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.282328] env[62460]: DEBUG oslo_vmware.api [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 994.282328] env[62460]: value = "task-1314010" [ 994.282328] env[62460]: _type = "Task" [ 994.282328] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.291965] env[62460]: DEBUG oslo_vmware.api [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314010, 'name': Rename_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.509171] env[62460]: DEBUG oslo_vmware.api [None req-9747b473-768f-4dbc-884d-4e0e4644fdc1 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1314004, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.584582] env[62460]: DEBUG oslo_concurrency.lockutils [req-26873b8c-4aea-435f-8236-117102c6ba21 req-6da18bae-1d66-4e1b-93e1-a48e3d7cba91 service nova] Releasing lock "refresh_cache-ddf1daf4-2772-48df-82c9-2192d012dedd" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.615289] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1fef2255-26dd-4541-a89a-93754111099b tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquiring lock "896f8ff4-e45f-4403-a727-03ee25e58609" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.615575] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1fef2255-26dd-4541-a89a-93754111099b tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lock "896f8ff4-e45f-4403-a727-03ee25e58609" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.705202] env[62460]: DEBUG oslo_vmware.api [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1314007, 'name': PowerOnVM_Task, 'duration_secs': 1.053992} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.705499] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 994.705711] env[62460]: INFO nova.compute.manager [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Took 8.93 seconds to spawn the instance on the hypervisor. [ 994.705900] env[62460]: DEBUG nova.compute.manager [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 994.706708] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-720a2d04-7352-4950-a7eb-55302a919645 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.732211] env[62460]: DEBUG oslo_vmware.api [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5285f332-fa0d-1dd7-e1f2-a7aaa8fda2b4, 'name': SearchDatastore_Task, 'duration_secs': 0.012403} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.733124] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ba9e9a6f-8f86-498b-aff7-6607a72b6c5a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.738019] env[62460]: DEBUG oslo_vmware.api [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Waiting for the task: (returnval){ [ 994.738019] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]5247e82e-ea40-7253-dae5-31769ff7d0a6" [ 994.738019] env[62460]: _type = "Task" [ 994.738019] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.745551] env[62460]: DEBUG oslo_vmware.api [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5247e82e-ea40-7253-dae5-31769ff7d0a6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.793355] env[62460]: DEBUG oslo_vmware.api [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314010, 'name': Rename_Task, 'duration_secs': 0.128248} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.794039] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 994.794170] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-61918846-c837-4277-b75f-1721da673893 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.802827] env[62460]: DEBUG oslo_vmware.api [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 994.802827] env[62460]: value = "task-1314011" [ 994.802827] env[62460]: _type = "Task" [ 994.802827] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.817291] env[62460]: DEBUG oslo_vmware.api [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314011, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.009583] env[62460]: DEBUG oslo_vmware.api [None req-9747b473-768f-4dbc-884d-4e0e4644fdc1 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1314004, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.119849] env[62460]: INFO nova.compute.manager [None req-1fef2255-26dd-4541-a89a-93754111099b tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Detaching volume c472b891-df79-428d-a905-5349e414e7f5 [ 995.157196] env[62460]: INFO nova.virt.block_device [None req-1fef2255-26dd-4541-a89a-93754111099b tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Attempting to driver detach volume c472b891-df79-428d-a905-5349e414e7f5 from mountpoint /dev/sdb [ 995.157519] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-1fef2255-26dd-4541-a89a-93754111099b tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Volume detach. Driver type: vmdk {{(pid=62460) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 995.157763] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-1fef2255-26dd-4541-a89a-93754111099b tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281244', 'volume_id': 'c472b891-df79-428d-a905-5349e414e7f5', 'name': 'volume-c472b891-df79-428d-a905-5349e414e7f5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '896f8ff4-e45f-4403-a727-03ee25e58609', 'attached_at': '', 'detached_at': '', 'volume_id': 'c472b891-df79-428d-a905-5349e414e7f5', 'serial': 'c472b891-df79-428d-a905-5349e414e7f5'} {{(pid=62460) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 995.158855] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2da815d1-2fc5-4efe-97b8-5b4abc484feb {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.182915] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c64b1289-5142-439e-85d5-0e8c0d4be179 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.190194] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-708077af-be58-484a-9fd8-fe9bbc3ba1ad {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.211379] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d74b1b0-b275-462e-9017-bbd4eb4e0c36 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.229575] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-1fef2255-26dd-4541-a89a-93754111099b tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] The volume has not been displaced from its original location: [datastore2] volume-c472b891-df79-428d-a905-5349e414e7f5/volume-c472b891-df79-428d-a905-5349e414e7f5.vmdk. No consolidation needed. {{(pid=62460) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 995.234895] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-1fef2255-26dd-4541-a89a-93754111099b tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Reconfiguring VM instance instance-0000004a to detach disk 2001 {{(pid=62460) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 995.236802] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-de81ac5a-a414-4fdc-9fa9-d39950e49d6a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.249403] env[62460]: INFO nova.compute.manager [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Took 16.56 seconds to build instance. [ 995.259191] env[62460]: DEBUG oslo_vmware.api [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5247e82e-ea40-7253-dae5-31769ff7d0a6, 'name': SearchDatastore_Task, 'duration_secs': 0.013421} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.260463] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 995.260689] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] ddf1daf4-2772-48df-82c9-2192d012dedd/ddf1daf4-2772-48df-82c9-2192d012dedd.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 995.261013] env[62460]: DEBUG oslo_vmware.api [None req-1fef2255-26dd-4541-a89a-93754111099b tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 995.261013] env[62460]: value = "task-1314012" [ 995.261013] env[62460]: _type = "Task" [ 995.261013] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.261739] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e21043a0-f9c3-432a-9629-27ed2913b4f3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.272349] env[62460]: DEBUG oslo_vmware.api [None req-1fef2255-26dd-4541-a89a-93754111099b tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314012, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.273487] env[62460]: DEBUG oslo_vmware.api [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Waiting for the task: (returnval){ [ 995.273487] env[62460]: value = "task-1314013" [ 995.273487] env[62460]: _type = "Task" [ 995.273487] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.282880] env[62460]: DEBUG oslo_vmware.api [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1314013, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.303998] env[62460]: DEBUG oslo_concurrency.lockutils [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "290500b5-d5b4-43b6-a843-20dacf561f4b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.304658] env[62460]: DEBUG oslo_concurrency.lockutils [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "290500b5-d5b4-43b6-a843-20dacf561f4b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.304658] env[62460]: DEBUG oslo_concurrency.lockutils [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "290500b5-d5b4-43b6-a843-20dacf561f4b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.304816] env[62460]: DEBUG oslo_concurrency.lockutils [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "290500b5-d5b4-43b6-a843-20dacf561f4b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.305031] env[62460]: DEBUG oslo_concurrency.lockutils [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "290500b5-d5b4-43b6-a843-20dacf561f4b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.310482] env[62460]: INFO nova.compute.manager [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Terminating instance [ 995.312112] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a4cfaad3-52ca-40c0-a929-9fe97bab5e86 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "0d39a304-6a49-4d87-bfa2-4ba3b383578a" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.312349] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a4cfaad3-52ca-40c0-a929-9fe97bab5e86 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "0d39a304-6a49-4d87-bfa2-4ba3b383578a" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.312546] env[62460]: DEBUG nova.compute.manager [None req-a4cfaad3-52ca-40c0-a929-9fe97bab5e86 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Going to confirm migration 3 {{(pid=62460) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 995.313963] env[62460]: DEBUG nova.compute.manager [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 995.314182] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 995.315379] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-819641c7-7278-4482-92e8-c2a769dff0f3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.321495] env[62460]: DEBUG oslo_vmware.api [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314011, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.325566] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 995.325804] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-060c624f-2641-4c17-b01d-848fb4fd4ae7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.332090] env[62460]: DEBUG oslo_vmware.api [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 995.332090] env[62460]: value = "task-1314014" [ 995.332090] env[62460]: _type = "Task" [ 995.332090] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.340264] env[62460]: DEBUG oslo_vmware.api [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1314014, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.509934] env[62460]: DEBUG oslo_vmware.api [None req-9747b473-768f-4dbc-884d-4e0e4644fdc1 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1314004, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.754801] env[62460]: DEBUG oslo_concurrency.lockutils [None req-8692faf4-bda8-4425-b2cc-91c39d5008a8 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "4c5e60fe-6366-4fb9-975f-52345de67478" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.088s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.774666] env[62460]: DEBUG oslo_vmware.api [None req-1fef2255-26dd-4541-a89a-93754111099b tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314012, 'name': ReconfigVM_Task, 'duration_secs': 0.256673} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.778666] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-1fef2255-26dd-4541-a89a-93754111099b tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Reconfigured VM instance instance-0000004a to detach disk 2001 {{(pid=62460) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 995.783362] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6a7d7d7e-9d1f-4055-8ba1-6a2cc810f1d6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.800248] env[62460]: DEBUG oslo_vmware.api [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1314013, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.523202} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.801405] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] ddf1daf4-2772-48df-82c9-2192d012dedd/ddf1daf4-2772-48df-82c9-2192d012dedd.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 995.801629] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 995.801938] env[62460]: DEBUG oslo_vmware.api [None req-1fef2255-26dd-4541-a89a-93754111099b tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 995.801938] env[62460]: value = "task-1314015" [ 995.801938] env[62460]: _type = "Task" [ 995.801938] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.802203] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9c3ec785-d43e-467e-bc4a-bf82f2d3ea8e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.812850] env[62460]: DEBUG oslo_vmware.api [None req-1fef2255-26dd-4541-a89a-93754111099b tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314015, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.814364] env[62460]: DEBUG oslo_vmware.api [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Waiting for the task: (returnval){ [ 995.814364] env[62460]: value = "task-1314016" [ 995.814364] env[62460]: _type = "Task" [ 995.814364] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.823445] env[62460]: DEBUG oslo_vmware.api [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314011, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.831324] env[62460]: DEBUG oslo_vmware.api [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1314016, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.840378] env[62460]: DEBUG oslo_vmware.api [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1314014, 'name': PowerOffVM_Task, 'duration_secs': 0.187334} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.840639] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 995.840855] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 995.841184] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a68b7cde-caca-4b22-83a5-64422f1a23ec {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.918666] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a4cfaad3-52ca-40c0-a929-9fe97bab5e86 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "refresh_cache-0d39a304-6a49-4d87-bfa2-4ba3b383578a" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.918877] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a4cfaad3-52ca-40c0-a929-9fe97bab5e86 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquired lock "refresh_cache-0d39a304-6a49-4d87-bfa2-4ba3b383578a" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.919084] env[62460]: DEBUG nova.network.neutron [None req-a4cfaad3-52ca-40c0-a929-9fe97bab5e86 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 995.919306] env[62460]: DEBUG nova.objects.instance [None req-a4cfaad3-52ca-40c0-a929-9fe97bab5e86 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lazy-loading 'info_cache' on Instance uuid 0d39a304-6a49-4d87-bfa2-4ba3b383578a {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 996.013343] env[62460]: DEBUG oslo_vmware.api [None req-9747b473-768f-4dbc-884d-4e0e4644fdc1 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1314004, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.222889] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 996.223274] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Deleting contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 996.223509] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Deleting the datastore file [datastore1] 290500b5-d5b4-43b6-a843-20dacf561f4b {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 996.223856] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-262ff97f-d97a-4cff-a2b9-b3d15c33c484 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.231498] env[62460]: DEBUG oslo_vmware.api [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for the task: (returnval){ [ 996.231498] env[62460]: value = "task-1314018" [ 996.231498] env[62460]: _type = "Task" [ 996.231498] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.241481] env[62460]: DEBUG oslo_vmware.api [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1314018, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.321058] env[62460]: DEBUG oslo_vmware.api [None req-1fef2255-26dd-4541-a89a-93754111099b tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314015, 'name': ReconfigVM_Task, 'duration_secs': 0.135064} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.324165] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-1fef2255-26dd-4541-a89a-93754111099b tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281244', 'volume_id': 'c472b891-df79-428d-a905-5349e414e7f5', 'name': 'volume-c472b891-df79-428d-a905-5349e414e7f5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '896f8ff4-e45f-4403-a727-03ee25e58609', 'attached_at': '', 'detached_at': '', 'volume_id': 'c472b891-df79-428d-a905-5349e414e7f5', 'serial': 'c472b891-df79-428d-a905-5349e414e7f5'} {{(pid=62460) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 996.326176] env[62460]: DEBUG oslo_vmware.api [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314011, 'name': PowerOnVM_Task} progress is 87%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.330791] env[62460]: DEBUG oslo_vmware.api [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1314016, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071117} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.331084] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 996.331880] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c66095b8-01f4-43bc-b0f8-73b25ee2bbbb {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.353249] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] ddf1daf4-2772-48df-82c9-2192d012dedd/ddf1daf4-2772-48df-82c9-2192d012dedd.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 996.353561] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-33cc70ae-c840-4577-b482-f82a497ea42b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.373773] env[62460]: DEBUG oslo_vmware.api [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Waiting for the task: (returnval){ [ 996.373773] env[62460]: value = "task-1314019" [ 996.373773] env[62460]: _type = "Task" [ 996.373773] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.382484] env[62460]: DEBUG oslo_vmware.api [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1314019, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.511439] env[62460]: DEBUG oslo_vmware.api [None req-9747b473-768f-4dbc-884d-4e0e4644fdc1 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1314004, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.548148] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c3cdf37a-673a-4643-8dfd-16498e31280d tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "4c5e60fe-6366-4fb9-975f-52345de67478" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.548443] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c3cdf37a-673a-4643-8dfd-16498e31280d tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "4c5e60fe-6366-4fb9-975f-52345de67478" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.548747] env[62460]: DEBUG nova.compute.manager [None req-c3cdf37a-673a-4643-8dfd-16498e31280d tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 996.549637] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ef31574-75a9-483c-8d96-24faf8ab8743 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.556012] env[62460]: DEBUG nova.compute.manager [None req-c3cdf37a-673a-4643-8dfd-16498e31280d tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62460) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 996.556605] env[62460]: DEBUG nova.objects.instance [None req-c3cdf37a-673a-4643-8dfd-16498e31280d tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lazy-loading 'flavor' on Instance uuid 4c5e60fe-6366-4fb9-975f-52345de67478 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 996.741971] env[62460]: DEBUG oslo_vmware.api [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Task: {'id': task-1314018, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.212618} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.742252] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 996.742478] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Deleted contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 996.742678] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 996.742866] env[62460]: INFO nova.compute.manager [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Took 1.43 seconds to destroy the instance on the hypervisor. [ 996.743135] env[62460]: DEBUG oslo.service.loopingcall [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 996.743343] env[62460]: DEBUG nova.compute.manager [-] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 996.743441] env[62460]: DEBUG nova.network.neutron [-] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 996.819830] env[62460]: DEBUG oslo_vmware.api [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314011, 'name': PowerOnVM_Task, 'duration_secs': 1.684414} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.820151] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 996.820383] env[62460]: INFO nova.compute.manager [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Took 8.76 seconds to spawn the instance on the hypervisor. [ 996.820574] env[62460]: DEBUG nova.compute.manager [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 996.821799] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-968bca7b-b164-42ee-9fd7-920433e4ed79 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.870217] env[62460]: DEBUG nova.objects.instance [None req-1fef2255-26dd-4541-a89a-93754111099b tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lazy-loading 'flavor' on Instance uuid 896f8ff4-e45f-4403-a727-03ee25e58609 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 996.883619] env[62460]: DEBUG oslo_vmware.api [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1314019, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.011328] env[62460]: DEBUG oslo_vmware.api [None req-9747b473-768f-4dbc-884d-4e0e4644fdc1 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1314004, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.022205] env[62460]: DEBUG nova.compute.manager [req-f1f3667c-3c1c-418e-acd6-e41dff2b5b1e req-238f0c44-b19f-40d3-ade0-096e4d034d49 service nova] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Received event network-vif-deleted-9f9c5b7b-922a-4387-b997-3f552dd85a50 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 997.022472] env[62460]: INFO nova.compute.manager [req-f1f3667c-3c1c-418e-acd6-e41dff2b5b1e req-238f0c44-b19f-40d3-ade0-096e4d034d49 service nova] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Neutron deleted interface 9f9c5b7b-922a-4387-b997-3f552dd85a50; detaching it from the instance and deleting it from the info cache [ 997.022563] env[62460]: DEBUG nova.network.neutron [req-f1f3667c-3c1c-418e-acd6-e41dff2b5b1e req-238f0c44-b19f-40d3-ade0-096e4d034d49 service nova] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 997.064291] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3cdf37a-673a-4643-8dfd-16498e31280d tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 997.065385] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0fffcc23-2043-4314-be67-4fb6fa2eee4d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.072072] env[62460]: DEBUG oslo_vmware.api [None req-c3cdf37a-673a-4643-8dfd-16498e31280d tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for the task: (returnval){ [ 997.072072] env[62460]: value = "task-1314020" [ 997.072072] env[62460]: _type = "Task" [ 997.072072] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.080181] env[62460]: DEBUG oslo_vmware.api [None req-c3cdf37a-673a-4643-8dfd-16498e31280d tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1314020, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.145557] env[62460]: DEBUG nova.network.neutron [None req-a4cfaad3-52ca-40c0-a929-9fe97bab5e86 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Updating instance_info_cache with network_info: [{"id": "786ef177-843f-44d0-8920-2f4332e29154", "address": "fa:16:3e:09:59:ab", "network": {"id": "3c49720c-94c9-4e2f-bbb7-4266cc97faee", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1650281790-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfa5a65b0f614e769de5b3aa77bf869e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap786ef177-84", "ovs_interfaceid": "786ef177-843f-44d0-8920-2f4332e29154", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 997.340933] env[62460]: INFO nova.compute.manager [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Took 16.04 seconds to build instance. [ 997.385467] env[62460]: DEBUG oslo_vmware.api [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1314019, 'name': ReconfigVM_Task, 'duration_secs': 0.57553} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.386329] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Reconfigured VM instance instance-00000064 to attach disk [datastore1] ddf1daf4-2772-48df-82c9-2192d012dedd/ddf1daf4-2772-48df-82c9-2192d012dedd.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 997.386994] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-85534a58-855a-4cbb-9060-ea39756db423 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.393870] env[62460]: DEBUG oslo_vmware.api [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Waiting for the task: (returnval){ [ 997.393870] env[62460]: value = "task-1314021" [ 997.393870] env[62460]: _type = "Task" [ 997.393870] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.401503] env[62460]: DEBUG oslo_vmware.api [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1314021, 'name': Rename_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.499449] env[62460]: DEBUG nova.network.neutron [-] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 997.512308] env[62460]: DEBUG oslo_vmware.api [None req-9747b473-768f-4dbc-884d-4e0e4644fdc1 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1314004, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.525466] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-365d721a-8714-4ab5-aada-494042d55117 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.534522] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-003090b3-a826-41d3-b230-755be48b8e9f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.563649] env[62460]: DEBUG nova.compute.manager [req-f1f3667c-3c1c-418e-acd6-e41dff2b5b1e req-238f0c44-b19f-40d3-ade0-096e4d034d49 service nova] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Detach interface failed, port_id=9f9c5b7b-922a-4387-b997-3f552dd85a50, reason: Instance 290500b5-d5b4-43b6-a843-20dacf561f4b could not be found. {{(pid=62460) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 997.584422] env[62460]: DEBUG oslo_vmware.api [None req-c3cdf37a-673a-4643-8dfd-16498e31280d tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1314020, 'name': PowerOffVM_Task, 'duration_secs': 0.417303} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.584422] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3cdf37a-673a-4643-8dfd-16498e31280d tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 997.584422] env[62460]: DEBUG nova.compute.manager [None req-c3cdf37a-673a-4643-8dfd-16498e31280d tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 997.584968] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f085d381-40bc-4684-8445-00f2e515ee2b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.648114] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a4cfaad3-52ca-40c0-a929-9fe97bab5e86 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Releasing lock "refresh_cache-0d39a304-6a49-4d87-bfa2-4ba3b383578a" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.648500] env[62460]: DEBUG nova.objects.instance [None req-a4cfaad3-52ca-40c0-a929-9fe97bab5e86 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lazy-loading 'migration_context' on Instance uuid 0d39a304-6a49-4d87-bfa2-4ba3b383578a {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 997.842928] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9bff0d1b-f465-4e8c-b7b3-25d94bf3be5e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "61f4666e-0588-4a00-a969-2349148caef2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.554s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.880695] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1fef2255-26dd-4541-a89a-93754111099b tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lock "896f8ff4-e45f-4403-a727-03ee25e58609" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.265s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.906575] env[62460]: DEBUG oslo_vmware.api [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1314021, 'name': Rename_Task, 'duration_secs': 0.210256} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.906778] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 997.907032] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d56957cd-aa26-47c4-9769-021d7f2eb3fc {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.912886] env[62460]: DEBUG oslo_vmware.api [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Waiting for the task: (returnval){ [ 997.912886] env[62460]: value = "task-1314022" [ 997.912886] env[62460]: _type = "Task" [ 997.912886] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.920708] env[62460]: DEBUG oslo_vmware.api [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1314022, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.002482] env[62460]: INFO nova.compute.manager [-] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Took 1.26 seconds to deallocate network for instance. [ 998.016028] env[62460]: DEBUG oslo_vmware.api [None req-9747b473-768f-4dbc-884d-4e0e4644fdc1 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1314004, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.097097] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c3cdf37a-673a-4643-8dfd-16498e31280d tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "4c5e60fe-6366-4fb9-975f-52345de67478" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.549s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.151216] env[62460]: DEBUG nova.objects.base [None req-a4cfaad3-52ca-40c0-a929-9fe97bab5e86 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Object Instance<0d39a304-6a49-4d87-bfa2-4ba3b383578a> lazy-loaded attributes: info_cache,migration_context {{(pid=62460) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 998.152354] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae4127b7-d876-4d51-8b4d-51b82655a172 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.174104] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ddc415b3-aa94-4ece-8b55-c051a97ec8cb {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.180736] env[62460]: DEBUG oslo_vmware.api [None req-a4cfaad3-52ca-40c0-a929-9fe97bab5e86 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 998.180736] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52bb939e-404a-c0e2-6275-d4957ab21c19" [ 998.180736] env[62460]: _type = "Task" [ 998.180736] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.188438] env[62460]: DEBUG oslo_vmware.api [None req-a4cfaad3-52ca-40c0-a929-9fe97bab5e86 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52bb939e-404a-c0e2-6275-d4957ab21c19, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.423731] env[62460]: DEBUG oslo_vmware.api [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1314022, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.515083] env[62460]: DEBUG oslo_concurrency.lockutils [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.515083] env[62460]: DEBUG oslo_concurrency.lockutils [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.515083] env[62460]: DEBUG oslo_concurrency.lockutils [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.516705] env[62460]: DEBUG oslo_vmware.api [None req-9747b473-768f-4dbc-884d-4e0e4644fdc1 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1314004, 'name': ReconfigVM_Task, 'duration_secs': 5.964755} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.517295] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9747b473-768f-4dbc-884d-4e0e4644fdc1 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Releasing lock "0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 998.517647] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-9747b473-768f-4dbc-884d-4e0e4644fdc1 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Reconfigured VM to detach interface {{(pid=62460) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 998.543149] env[62460]: INFO nova.scheduler.client.report [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Deleted allocations for instance 290500b5-d5b4-43b6-a843-20dacf561f4b [ 998.692873] env[62460]: DEBUG oslo_vmware.api [None req-a4cfaad3-52ca-40c0-a929-9fe97bab5e86 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52bb939e-404a-c0e2-6275-d4957ab21c19, 'name': SearchDatastore_Task, 'duration_secs': 0.006799} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.692873] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a4cfaad3-52ca-40c0-a929-9fe97bab5e86 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.693128] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a4cfaad3-52ca-40c0-a929-9fe97bab5e86 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.880185] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7c85a888-2a2d-4bed-94e6-874a37ef74e2 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquiring lock "896f8ff4-e45f-4403-a727-03ee25e58609" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.880505] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7c85a888-2a2d-4bed-94e6-874a37ef74e2 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lock "896f8ff4-e45f-4403-a727-03ee25e58609" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.880762] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7c85a888-2a2d-4bed-94e6-874a37ef74e2 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquiring lock "896f8ff4-e45f-4403-a727-03ee25e58609-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.880975] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7c85a888-2a2d-4bed-94e6-874a37ef74e2 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lock "896f8ff4-e45f-4403-a727-03ee25e58609-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.881205] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7c85a888-2a2d-4bed-94e6-874a37ef74e2 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lock "896f8ff4-e45f-4403-a727-03ee25e58609-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.883538] env[62460]: INFO nova.compute.manager [None req-7c85a888-2a2d-4bed-94e6-874a37ef74e2 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Terminating instance [ 998.885703] env[62460]: DEBUG nova.compute.manager [None req-7c85a888-2a2d-4bed-94e6-874a37ef74e2 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 998.886006] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7c85a888-2a2d-4bed-94e6-874a37ef74e2 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 998.886961] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-414420f0-b6e5-47a5-980a-fb70b284f70a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.896540] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c85a888-2a2d-4bed-94e6-874a37ef74e2 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 998.896540] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-683c9abe-798a-42fd-9c2f-3253d54f427e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.903992] env[62460]: DEBUG oslo_vmware.api [None req-7c85a888-2a2d-4bed-94e6-874a37ef74e2 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 998.903992] env[62460]: value = "task-1314023" [ 998.903992] env[62460]: _type = "Task" [ 998.903992] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.912284] env[62460]: DEBUG oslo_vmware.api [None req-7c85a888-2a2d-4bed-94e6-874a37ef74e2 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314023, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.924453] env[62460]: DEBUG oslo_vmware.api [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1314022, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.049024] env[62460]: DEBUG nova.compute.manager [req-3cc1d058-9132-4273-9df0-683a3c73a85e req-e8b77439-2933-4d7d-a83f-0f280fe0095e service nova] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Received event network-changed-345c87ec-2bb1-420f-bc8c-845a795baad5 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 999.049263] env[62460]: DEBUG nova.compute.manager [req-3cc1d058-9132-4273-9df0-683a3c73a85e req-e8b77439-2933-4d7d-a83f-0f280fe0095e service nova] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Refreshing instance network info cache due to event network-changed-345c87ec-2bb1-420f-bc8c-845a795baad5. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 999.049489] env[62460]: DEBUG oslo_concurrency.lockutils [req-3cc1d058-9132-4273-9df0-683a3c73a85e req-e8b77439-2933-4d7d-a83f-0f280fe0095e service nova] Acquiring lock "refresh_cache-61f4666e-0588-4a00-a969-2349148caef2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 999.049641] env[62460]: DEBUG oslo_concurrency.lockutils [req-3cc1d058-9132-4273-9df0-683a3c73a85e req-e8b77439-2933-4d7d-a83f-0f280fe0095e service nova] Acquired lock "refresh_cache-61f4666e-0588-4a00-a969-2349148caef2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.049808] env[62460]: DEBUG nova.network.neutron [req-3cc1d058-9132-4273-9df0-683a3c73a85e req-e8b77439-2933-4d7d-a83f-0f280fe0095e service nova] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Refreshing network info cache for port 345c87ec-2bb1-420f-bc8c-845a795baad5 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 999.053870] env[62460]: DEBUG oslo_concurrency.lockutils [None req-96229ff9-9cd4-46f9-9e0a-140dc637477b tempest-DeleteServersTestJSON-2091268211 tempest-DeleteServersTestJSON-2091268211-project-member] Lock "290500b5-d5b4-43b6-a843-20dacf561f4b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.750s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.285171] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3e8a5cc8-9c83-40a7-8fc2-b8f324357541 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "4c5e60fe-6366-4fb9-975f-52345de67478" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.285549] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3e8a5cc8-9c83-40a7-8fc2-b8f324357541 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "4c5e60fe-6366-4fb9-975f-52345de67478" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.285710] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3e8a5cc8-9c83-40a7-8fc2-b8f324357541 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "4c5e60fe-6366-4fb9-975f-52345de67478-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.285916] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3e8a5cc8-9c83-40a7-8fc2-b8f324357541 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "4c5e60fe-6366-4fb9-975f-52345de67478-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.286137] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3e8a5cc8-9c83-40a7-8fc2-b8f324357541 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "4c5e60fe-6366-4fb9-975f-52345de67478-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.290702] env[62460]: INFO nova.compute.manager [None req-3e8a5cc8-9c83-40a7-8fc2-b8f324357541 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Terminating instance [ 999.292901] env[62460]: DEBUG nova.compute.manager [None req-3e8a5cc8-9c83-40a7-8fc2-b8f324357541 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 999.293130] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-3e8a5cc8-9c83-40a7-8fc2-b8f324357541 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 999.293971] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b4c8578-a2dd-4b42-909c-f7630f92881d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.301778] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-3e8a5cc8-9c83-40a7-8fc2-b8f324357541 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 999.303051] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7a1ac70f-2ebe-40d0-8b2c-760dd8f7011a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.330389] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fab24aec-af8d-49d3-9367-281b0f23b136 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.338480] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab7de04d-aa51-460c-8fc7-7075a4f366c8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.372256] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6843f811-ed3d-4485-9e05-eab894cbff75 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.374880] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-3e8a5cc8-9c83-40a7-8fc2-b8f324357541 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 999.375048] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-3e8a5cc8-9c83-40a7-8fc2-b8f324357541 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Deleting contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 999.375308] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e8a5cc8-9c83-40a7-8fc2-b8f324357541 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Deleting the datastore file [datastore2] 4c5e60fe-6366-4fb9-975f-52345de67478 {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 999.376672] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ce26c702-aed6-4149-9150-c917334d8c6c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.384308] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b5f9f93-bbe6-4b26-949e-0978addc3fc4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.388103] env[62460]: DEBUG oslo_vmware.api [None req-3e8a5cc8-9c83-40a7-8fc2-b8f324357541 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for the task: (returnval){ [ 999.388103] env[62460]: value = "task-1314025" [ 999.388103] env[62460]: _type = "Task" [ 999.388103] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.400140] env[62460]: DEBUG nova.compute.provider_tree [None req-a4cfaad3-52ca-40c0-a929-9fe97bab5e86 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 999.406521] env[62460]: DEBUG oslo_vmware.api [None req-3e8a5cc8-9c83-40a7-8fc2-b8f324357541 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1314025, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.415351] env[62460]: DEBUG oslo_vmware.api [None req-7c85a888-2a2d-4bed-94e6-874a37ef74e2 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314023, 'name': PowerOffVM_Task, 'duration_secs': 0.372347} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.418376] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c85a888-2a2d-4bed-94e6-874a37ef74e2 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 999.418567] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7c85a888-2a2d-4bed-94e6-874a37ef74e2 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 999.418832] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-619a1489-9e68-467c-9b34-bad6e7d6084c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.425382] env[62460]: DEBUG oslo_vmware.api [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1314022, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.485044] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7c85a888-2a2d-4bed-94e6-874a37ef74e2 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 999.485044] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7c85a888-2a2d-4bed-94e6-874a37ef74e2 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Deleting contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 999.485044] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-7c85a888-2a2d-4bed-94e6-874a37ef74e2 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Deleting the datastore file [datastore2] 896f8ff4-e45f-4403-a727-03ee25e58609 {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 999.485044] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1d25f859-22a8-4ee3-9193-89fbadcf4b1b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.492513] env[62460]: DEBUG oslo_vmware.api [None req-7c85a888-2a2d-4bed-94e6-874a37ef74e2 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 999.492513] env[62460]: value = "task-1314027" [ 999.492513] env[62460]: _type = "Task" [ 999.492513] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.501944] env[62460]: DEBUG oslo_vmware.api [None req-7c85a888-2a2d-4bed-94e6-874a37ef74e2 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314027, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.879407] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9747b473-768f-4dbc-884d-4e0e4644fdc1 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "refresh_cache-0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 999.879613] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9747b473-768f-4dbc-884d-4e0e4644fdc1 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquired lock "refresh_cache-0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.879787] env[62460]: DEBUG nova.network.neutron [None req-9747b473-768f-4dbc-884d-4e0e4644fdc1 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 999.900168] env[62460]: DEBUG oslo_vmware.api [None req-3e8a5cc8-9c83-40a7-8fc2-b8f324357541 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Task: {'id': task-1314025, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.17347} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.901507] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e8a5cc8-9c83-40a7-8fc2-b8f324357541 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 999.901507] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-3e8a5cc8-9c83-40a7-8fc2-b8f324357541 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Deleted contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 999.901507] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-3e8a5cc8-9c83-40a7-8fc2-b8f324357541 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 999.901507] env[62460]: INFO nova.compute.manager [None req-3e8a5cc8-9c83-40a7-8fc2-b8f324357541 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Took 0.61 seconds to destroy the instance on the hypervisor. [ 999.901507] env[62460]: DEBUG oslo.service.loopingcall [None req-3e8a5cc8-9c83-40a7-8fc2-b8f324357541 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 999.901805] env[62460]: DEBUG nova.compute.manager [-] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 999.901805] env[62460]: DEBUG nova.network.neutron [-] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 999.904954] env[62460]: DEBUG nova.scheduler.client.report [None req-a4cfaad3-52ca-40c0-a929-9fe97bab5e86 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 999.930029] env[62460]: DEBUG oslo_vmware.api [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1314022, 'name': PowerOnVM_Task, 'duration_secs': 1.750516} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.930029] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 999.930029] env[62460]: INFO nova.compute.manager [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Took 9.61 seconds to spawn the instance on the hypervisor. [ 999.930029] env[62460]: DEBUG nova.compute.manager [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 999.930029] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c829fa70-de96-49a0-86ce-6b41b7f0d195 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.940212] env[62460]: DEBUG nova.network.neutron [req-3cc1d058-9132-4273-9df0-683a3c73a85e req-e8b77439-2933-4d7d-a83f-0f280fe0095e service nova] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Updated VIF entry in instance network info cache for port 345c87ec-2bb1-420f-bc8c-845a795baad5. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 999.944019] env[62460]: DEBUG nova.network.neutron [req-3cc1d058-9132-4273-9df0-683a3c73a85e req-e8b77439-2933-4d7d-a83f-0f280fe0095e service nova] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Updating instance_info_cache with network_info: [{"id": "345c87ec-2bb1-420f-bc8c-845a795baad5", "address": "fa:16:3e:43:36:b0", "network": {"id": "1dfc278d-e89e-4caf-9cfb-c302106cf974", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-868970047-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eee8bf23869e4680aec16e37f972175f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap345c87ec-2b", "ovs_interfaceid": "345c87ec-2bb1-420f-bc8c-845a795baad5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1000.003798] env[62460]: DEBUG oslo_vmware.api [None req-7c85a888-2a2d-4bed-94e6-874a37ef74e2 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314027, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.289303} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.003798] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-7c85a888-2a2d-4bed-94e6-874a37ef74e2 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1000.003798] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7c85a888-2a2d-4bed-94e6-874a37ef74e2 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Deleted contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1000.004083] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7c85a888-2a2d-4bed-94e6-874a37ef74e2 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1000.004151] env[62460]: INFO nova.compute.manager [None req-7c85a888-2a2d-4bed-94e6-874a37ef74e2 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1000.004411] env[62460]: DEBUG oslo.service.loopingcall [None req-7c85a888-2a2d-4bed-94e6-874a37ef74e2 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1000.004624] env[62460]: DEBUG nova.compute.manager [-] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1000.004721] env[62460]: DEBUG nova.network.neutron [-] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1000.401321] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b7594f06-69f8-4a0a-b7cc-5eecc483a3d6 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.401619] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b7594f06-69f8-4a0a-b7cc-5eecc483a3d6 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.401839] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b7594f06-69f8-4a0a-b7cc-5eecc483a3d6 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.402045] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b7594f06-69f8-4a0a-b7cc-5eecc483a3d6 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.402230] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b7594f06-69f8-4a0a-b7cc-5eecc483a3d6 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.404908] env[62460]: INFO nova.compute.manager [None req-b7594f06-69f8-4a0a-b7cc-5eecc483a3d6 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Terminating instance [ 1000.409047] env[62460]: DEBUG nova.compute.manager [None req-b7594f06-69f8-4a0a-b7cc-5eecc483a3d6 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1000.409269] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-b7594f06-69f8-4a0a-b7cc-5eecc483a3d6 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1000.418604] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec9f61d7-0529-4f83-a8cd-2e05f0c231b5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.428368] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7594f06-69f8-4a0a-b7cc-5eecc483a3d6 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1000.428728] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f555a0b7-6879-4c3b-b913-d3a66ec8458c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.435467] env[62460]: DEBUG oslo_vmware.api [None req-b7594f06-69f8-4a0a-b7cc-5eecc483a3d6 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Waiting for the task: (returnval){ [ 1000.435467] env[62460]: value = "task-1314028" [ 1000.435467] env[62460]: _type = "Task" [ 1000.435467] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.443950] env[62460]: DEBUG oslo_concurrency.lockutils [req-3cc1d058-9132-4273-9df0-683a3c73a85e req-e8b77439-2933-4d7d-a83f-0f280fe0095e service nova] Releasing lock "refresh_cache-61f4666e-0588-4a00-a969-2349148caef2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1000.451048] env[62460]: INFO nova.compute.manager [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Took 15.14 seconds to build instance. [ 1000.452543] env[62460]: DEBUG oslo_vmware.api [None req-b7594f06-69f8-4a0a-b7cc-5eecc483a3d6 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1314028, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.735813] env[62460]: DEBUG nova.network.neutron [-] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1000.813179] env[62460]: DEBUG nova.compute.manager [req-dcf6d0d5-dc1d-4e1c-9581-f97b8024e448 req-64af3a46-2013-4f7c-84eb-37ae0cb50020 service nova] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Received event network-vif-deleted-72a87ddf-0585-429a-b9de-d73bcad42cd1 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1000.813179] env[62460]: INFO nova.compute.manager [req-dcf6d0d5-dc1d-4e1c-9581-f97b8024e448 req-64af3a46-2013-4f7c-84eb-37ae0cb50020 service nova] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Neutron deleted interface 72a87ddf-0585-429a-b9de-d73bcad42cd1; detaching it from the instance and deleting it from the info cache [ 1000.813179] env[62460]: DEBUG nova.network.neutron [req-dcf6d0d5-dc1d-4e1c-9581-f97b8024e448 req-64af3a46-2013-4f7c-84eb-37ae0cb50020 service nova] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1000.856617] env[62460]: INFO nova.network.neutron [None req-9747b473-768f-4dbc-884d-4e0e4644fdc1 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Port 50d6a0c4-3681-4c1a-b42f-2987ee7a3bfb from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1000.856842] env[62460]: DEBUG nova.network.neutron [None req-9747b473-768f-4dbc-884d-4e0e4644fdc1 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Updating instance_info_cache with network_info: [{"id": "90a7d774-4d76-4525-8e6d-8c08fdb93fb4", "address": "fa:16:3e:5b:1c:b7", "network": {"id": "d3136e32-ad55-4b73-835c-8fa5f0480767", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1511414968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53fa06008e9f43488362895e7a143700", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "054fcd1e-638e-425a-a1de-78cb188ae026", "external-id": "nsx-vlan-transportzone-658", "segmentation_id": 658, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90a7d774-4d", "ovs_interfaceid": "90a7d774-4d76-4525-8e6d-8c08fdb93fb4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1000.924487] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a4cfaad3-52ca-40c0-a929-9fe97bab5e86 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.231s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.947121] env[62460]: DEBUG oslo_vmware.api [None req-b7594f06-69f8-4a0a-b7cc-5eecc483a3d6 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1314028, 'name': PowerOffVM_Task, 'duration_secs': 0.284134} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.947530] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7594f06-69f8-4a0a-b7cc-5eecc483a3d6 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1000.947692] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-b7594f06-69f8-4a0a-b7cc-5eecc483a3d6 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1000.948229] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a28e6af9-1d53-42f9-8168-d0f5e14bb987 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.953174] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2df67c85-a7d6-4a36-94e8-6d54aad5b5d7 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lock "ddf1daf4-2772-48df-82c9-2192d012dedd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.651s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.021804] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-b7594f06-69f8-4a0a-b7cc-5eecc483a3d6 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1001.022038] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-b7594f06-69f8-4a0a-b7cc-5eecc483a3d6 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Deleting contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1001.022248] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7594f06-69f8-4a0a-b7cc-5eecc483a3d6 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Deleting the datastore file [datastore1] 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1001.022552] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-30ece6c9-1ffc-4d6d-999b-b7d68893cf12 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.028627] env[62460]: DEBUG oslo_vmware.api [None req-b7594f06-69f8-4a0a-b7cc-5eecc483a3d6 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Waiting for the task: (returnval){ [ 1001.028627] env[62460]: value = "task-1314030" [ 1001.028627] env[62460]: _type = "Task" [ 1001.028627] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.037576] env[62460]: DEBUG oslo_vmware.api [None req-b7594f06-69f8-4a0a-b7cc-5eecc483a3d6 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1314030, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.079587] env[62460]: DEBUG nova.compute.manager [req-c8a216a1-cac6-4e84-aac4-dfb08407f71d req-9401d033-b708-4b77-a186-db454d532e78 service nova] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Received event network-vif-deleted-9dbf899d-26a8-45cd-994c-1913dd0b91d2 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1001.079840] env[62460]: DEBUG nova.compute.manager [req-c8a216a1-cac6-4e84-aac4-dfb08407f71d req-9401d033-b708-4b77-a186-db454d532e78 service nova] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Received event network-changed-5fe76c2c-4427-44ee-b05d-f0e6fbbd874e {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1001.079995] env[62460]: DEBUG nova.compute.manager [req-c8a216a1-cac6-4e84-aac4-dfb08407f71d req-9401d033-b708-4b77-a186-db454d532e78 service nova] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Refreshing instance network info cache due to event network-changed-5fe76c2c-4427-44ee-b05d-f0e6fbbd874e. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1001.080326] env[62460]: DEBUG oslo_concurrency.lockutils [req-c8a216a1-cac6-4e84-aac4-dfb08407f71d req-9401d033-b708-4b77-a186-db454d532e78 service nova] Acquiring lock "refresh_cache-ddf1daf4-2772-48df-82c9-2192d012dedd" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1001.080483] env[62460]: DEBUG oslo_concurrency.lockutils [req-c8a216a1-cac6-4e84-aac4-dfb08407f71d req-9401d033-b708-4b77-a186-db454d532e78 service nova] Acquired lock "refresh_cache-ddf1daf4-2772-48df-82c9-2192d012dedd" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.080656] env[62460]: DEBUG nova.network.neutron [req-c8a216a1-cac6-4e84-aac4-dfb08407f71d req-9401d033-b708-4b77-a186-db454d532e78 service nova] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Refreshing network info cache for port 5fe76c2c-4427-44ee-b05d-f0e6fbbd874e {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1001.238757] env[62460]: INFO nova.compute.manager [-] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Took 1.34 seconds to deallocate network for instance. [ 1001.284691] env[62460]: DEBUG nova.network.neutron [-] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.319503] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6a8cb9ec-753f-4575-a78e-fd536ea06bce {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.328810] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b7d83e9-beda-4c94-b370-130e5e5808ff {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.355780] env[62460]: DEBUG nova.compute.manager [req-dcf6d0d5-dc1d-4e1c-9581-f97b8024e448 req-64af3a46-2013-4f7c-84eb-37ae0cb50020 service nova] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Detach interface failed, port_id=72a87ddf-0585-429a-b9de-d73bcad42cd1, reason: Instance 896f8ff4-e45f-4403-a727-03ee25e58609 could not be found. {{(pid=62460) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1001.363275] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9747b473-768f-4dbc-884d-4e0e4644fdc1 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Releasing lock "refresh_cache-0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.488588] env[62460]: INFO nova.scheduler.client.report [None req-a4cfaad3-52ca-40c0-a929-9fe97bab5e86 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Deleted allocation for migration f48aedf1-a18c-4d31-9f53-6e452541bb4d [ 1001.539323] env[62460]: DEBUG oslo_vmware.api [None req-b7594f06-69f8-4a0a-b7cc-5eecc483a3d6 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1314030, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.18063} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.539659] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7594f06-69f8-4a0a-b7cc-5eecc483a3d6 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1001.539801] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-b7594f06-69f8-4a0a-b7cc-5eecc483a3d6 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Deleted contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1001.540037] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-b7594f06-69f8-4a0a-b7cc-5eecc483a3d6 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1001.540188] env[62460]: INFO nova.compute.manager [None req-b7594f06-69f8-4a0a-b7cc-5eecc483a3d6 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1001.540449] env[62460]: DEBUG oslo.service.loopingcall [None req-b7594f06-69f8-4a0a-b7cc-5eecc483a3d6 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1001.540649] env[62460]: DEBUG nova.compute.manager [-] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1001.540760] env[62460]: DEBUG nova.network.neutron [-] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1001.746344] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3e8a5cc8-9c83-40a7-8fc2-b8f324357541 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.746796] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3e8a5cc8-9c83-40a7-8fc2-b8f324357541 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.747187] env[62460]: DEBUG nova.objects.instance [None req-3e8a5cc8-9c83-40a7-8fc2-b8f324357541 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lazy-loading 'resources' on Instance uuid 4c5e60fe-6366-4fb9-975f-52345de67478 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1001.787440] env[62460]: INFO nova.compute.manager [-] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Took 1.78 seconds to deallocate network for instance. [ 1001.863680] env[62460]: DEBUG oslo_concurrency.lockutils [None req-9747b473-768f-4dbc-884d-4e0e4644fdc1 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "interface-0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea-50d6a0c4-3681-4c1a-b42f-2987ee7a3bfb" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.937s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.993992] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a4cfaad3-52ca-40c0-a929-9fe97bab5e86 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "0d39a304-6a49-4d87-bfa2-4ba3b383578a" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.681s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.093136] env[62460]: DEBUG nova.network.neutron [req-c8a216a1-cac6-4e84-aac4-dfb08407f71d req-9401d033-b708-4b77-a186-db454d532e78 service nova] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Updated VIF entry in instance network info cache for port 5fe76c2c-4427-44ee-b05d-f0e6fbbd874e. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1002.093578] env[62460]: DEBUG nova.network.neutron [req-c8a216a1-cac6-4e84-aac4-dfb08407f71d req-9401d033-b708-4b77-a186-db454d532e78 service nova] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Updating instance_info_cache with network_info: [{"id": "5fe76c2c-4427-44ee-b05d-f0e6fbbd874e", "address": "fa:16:3e:d6:e3:16", "network": {"id": "4cb03486-3bea-4c6a-a019-37486f271107", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-2094026403-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.182", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfc48e301fb04231b88ec28a062bda4e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c7d2575f-b92f-44ec-a863-634cb76631a2", "external-id": "nsx-vlan-transportzone-794", "segmentation_id": 794, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5fe76c2c-44", "ovs_interfaceid": "5fe76c2c-4427-44ee-b05d-f0e6fbbd874e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.297708] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7c85a888-2a2d-4bed-94e6-874a37ef74e2 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.382373] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4742779-5db7-4fbc-969a-b87f60f7d1a7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.391090] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ae15275-d7bb-4983-92db-5ae5571e90dc {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.425748] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00eb2f93-9547-4951-a19b-139cc0a17f44 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.433308] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5598ed0-42c3-4b83-a234-67890f52e864 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.446588] env[62460]: DEBUG nova.compute.provider_tree [None req-3e8a5cc8-9c83-40a7-8fc2-b8f324357541 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1002.596687] env[62460]: DEBUG oslo_concurrency.lockutils [req-c8a216a1-cac6-4e84-aac4-dfb08407f71d req-9401d033-b708-4b77-a186-db454d532e78 service nova] Releasing lock "refresh_cache-ddf1daf4-2772-48df-82c9-2192d012dedd" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.951031] env[62460]: DEBUG nova.scheduler.client.report [None req-3e8a5cc8-9c83-40a7-8fc2-b8f324357541 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1003.039279] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b9b15fa3-0ba6-440c-9cb3-48dc854f853f tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "0d39a304-6a49-4d87-bfa2-4ba3b383578a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.039582] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b9b15fa3-0ba6-440c-9cb3-48dc854f853f tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "0d39a304-6a49-4d87-bfa2-4ba3b383578a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.039798] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b9b15fa3-0ba6-440c-9cb3-48dc854f853f tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "0d39a304-6a49-4d87-bfa2-4ba3b383578a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.039991] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b9b15fa3-0ba6-440c-9cb3-48dc854f853f tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "0d39a304-6a49-4d87-bfa2-4ba3b383578a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.040240] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b9b15fa3-0ba6-440c-9cb3-48dc854f853f tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "0d39a304-6a49-4d87-bfa2-4ba3b383578a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.043015] env[62460]: INFO nova.compute.manager [None req-b9b15fa3-0ba6-440c-9cb3-48dc854f853f tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Terminating instance [ 1003.045100] env[62460]: DEBUG nova.compute.manager [None req-b9b15fa3-0ba6-440c-9cb3-48dc854f853f tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1003.045327] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-b9b15fa3-0ba6-440c-9cb3-48dc854f853f tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1003.046199] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1182efbb-542d-4a13-8944-eab67fa4c192 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.055443] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9b15fa3-0ba6-440c-9cb3-48dc854f853f tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1003.055657] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3a0a7742-3c84-4357-b80e-946d480c8c57 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.062230] env[62460]: DEBUG oslo_vmware.api [None req-b9b15fa3-0ba6-440c-9cb3-48dc854f853f tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 1003.062230] env[62460]: value = "task-1314031" [ 1003.062230] env[62460]: _type = "Task" [ 1003.062230] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.069665] env[62460]: DEBUG oslo_vmware.api [None req-b9b15fa3-0ba6-440c-9cb3-48dc854f853f tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1314031, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.110804] env[62460]: DEBUG nova.compute.manager [req-2f462155-31a1-4769-9ac4-6549c28ba180 req-4250110c-be9c-4660-837a-5f414c9a2867 service nova] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Received event network-vif-deleted-90a7d774-4d76-4525-8e6d-8c08fdb93fb4 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1003.111028] env[62460]: INFO nova.compute.manager [req-2f462155-31a1-4769-9ac4-6549c28ba180 req-4250110c-be9c-4660-837a-5f414c9a2867 service nova] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Neutron deleted interface 90a7d774-4d76-4525-8e6d-8c08fdb93fb4; detaching it from the instance and deleting it from the info cache [ 1003.111218] env[62460]: DEBUG nova.network.neutron [req-2f462155-31a1-4769-9ac4-6549c28ba180 req-4250110c-be9c-4660-837a-5f414c9a2867 service nova] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.142521] env[62460]: DEBUG nova.network.neutron [-] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.462447] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3e8a5cc8-9c83-40a7-8fc2-b8f324357541 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.715s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.465283] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7c85a888-2a2d-4bed-94e6-874a37ef74e2 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.167s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.465283] env[62460]: DEBUG nova.objects.instance [None req-7c85a888-2a2d-4bed-94e6-874a37ef74e2 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lazy-loading 'resources' on Instance uuid 896f8ff4-e45f-4403-a727-03ee25e58609 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1003.486039] env[62460]: INFO nova.scheduler.client.report [None req-3e8a5cc8-9c83-40a7-8fc2-b8f324357541 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Deleted allocations for instance 4c5e60fe-6366-4fb9-975f-52345de67478 [ 1003.512251] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "37e76fbf-fc84-4f63-ba7f-3515ccb376c1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.512491] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "37e76fbf-fc84-4f63-ba7f-3515ccb376c1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.572669] env[62460]: DEBUG oslo_vmware.api [None req-b9b15fa3-0ba6-440c-9cb3-48dc854f853f tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1314031, 'name': PowerOffVM_Task, 'duration_secs': 0.227695} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.572945] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9b15fa3-0ba6-440c-9cb3-48dc854f853f tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1003.573145] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-b9b15fa3-0ba6-440c-9cb3-48dc854f853f tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1003.573403] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f58c489e-f1a3-48ec-9005-5f6a363684a0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.614533] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-92a32850-4dcd-4b45-97d0-992cc9756fff {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.623507] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4ebb650-55ca-4fb1-896f-54edb6f93fda {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.647467] env[62460]: INFO nova.compute.manager [-] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Took 2.11 seconds to deallocate network for instance. [ 1003.647841] env[62460]: DEBUG nova.compute.manager [req-2f462155-31a1-4769-9ac4-6549c28ba180 req-4250110c-be9c-4660-837a-5f414c9a2867 service nova] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Detach interface failed, port_id=90a7d774-4d76-4525-8e6d-8c08fdb93fb4, reason: Instance 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea could not be found. {{(pid=62460) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1003.852423] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-b9b15fa3-0ba6-440c-9cb3-48dc854f853f tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1003.852696] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-b9b15fa3-0ba6-440c-9cb3-48dc854f853f tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Deleting contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1003.852906] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9b15fa3-0ba6-440c-9cb3-48dc854f853f tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Deleting the datastore file [datastore1] 0d39a304-6a49-4d87-bfa2-4ba3b383578a {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1003.853196] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-06519926-73b0-4d6e-8009-0d7269dbde40 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.859498] env[62460]: DEBUG oslo_vmware.api [None req-b9b15fa3-0ba6-440c-9cb3-48dc854f853f tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 1003.859498] env[62460]: value = "task-1314033" [ 1003.859498] env[62460]: _type = "Task" [ 1003.859498] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.867267] env[62460]: DEBUG oslo_vmware.api [None req-b9b15fa3-0ba6-440c-9cb3-48dc854f853f tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1314033, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.993281] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3e8a5cc8-9c83-40a7-8fc2-b8f324357541 tempest-ServersTestJSON-944294671 tempest-ServersTestJSON-944294671-project-member] Lock "4c5e60fe-6366-4fb9-975f-52345de67478" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.708s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.014654] env[62460]: DEBUG nova.compute.manager [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1004.080046] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f76c2a2-55b5-414a-9ad0-ef4cdc695b6b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.088187] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32ed97db-6977-4416-9c28-428175dd0eaa {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.118777] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9ea7e99-2b1c-4396-a939-33fb3f3995f1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.126487] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3014627c-081a-48e5-b4b9-0f04919c4414 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.139827] env[62460]: DEBUG nova.compute.provider_tree [None req-7c85a888-2a2d-4bed-94e6-874a37ef74e2 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1004.153719] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b7594f06-69f8-4a0a-b7cc-5eecc483a3d6 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.370558] env[62460]: DEBUG oslo_vmware.api [None req-b9b15fa3-0ba6-440c-9cb3-48dc854f853f tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1314033, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.179603} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.370826] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9b15fa3-0ba6-440c-9cb3-48dc854f853f tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1004.371082] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-b9b15fa3-0ba6-440c-9cb3-48dc854f853f tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Deleted contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1004.371311] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-b9b15fa3-0ba6-440c-9cb3-48dc854f853f tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1004.372022] env[62460]: INFO nova.compute.manager [None req-b9b15fa3-0ba6-440c-9cb3-48dc854f853f tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Took 1.33 seconds to destroy the instance on the hypervisor. [ 1004.372022] env[62460]: DEBUG oslo.service.loopingcall [None req-b9b15fa3-0ba6-440c-9cb3-48dc854f853f tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1004.372022] env[62460]: DEBUG nova.compute.manager [-] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1004.372230] env[62460]: DEBUG nova.network.neutron [-] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1004.544963] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.642637] env[62460]: DEBUG nova.scheduler.client.report [None req-7c85a888-2a2d-4bed-94e6-874a37ef74e2 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1005.129378] env[62460]: DEBUG nova.network.neutron [-] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1005.142965] env[62460]: DEBUG nova.compute.manager [req-0c0d2bfb-cc75-41e4-9d2f-c6c63274388c req-15402eb5-c670-4ea0-b014-8922bd713a9b service nova] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Received event network-vif-deleted-786ef177-843f-44d0-8920-2f4332e29154 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1005.143179] env[62460]: INFO nova.compute.manager [req-0c0d2bfb-cc75-41e4-9d2f-c6c63274388c req-15402eb5-c670-4ea0-b014-8922bd713a9b service nova] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Neutron deleted interface 786ef177-843f-44d0-8920-2f4332e29154; detaching it from the instance and deleting it from the info cache [ 1005.143406] env[62460]: DEBUG nova.network.neutron [req-0c0d2bfb-cc75-41e4-9d2f-c6c63274388c req-15402eb5-c670-4ea0-b014-8922bd713a9b service nova] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1005.149203] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7c85a888-2a2d-4bed-94e6-874a37ef74e2 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.685s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.151850] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b7594f06-69f8-4a0a-b7cc-5eecc483a3d6 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.998s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.152939] env[62460]: DEBUG nova.objects.instance [None req-b7594f06-69f8-4a0a-b7cc-5eecc483a3d6 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lazy-loading 'resources' on Instance uuid 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1005.182508] env[62460]: INFO nova.scheduler.client.report [None req-7c85a888-2a2d-4bed-94e6-874a37ef74e2 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Deleted allocations for instance 896f8ff4-e45f-4403-a727-03ee25e58609 [ 1005.637974] env[62460]: INFO nova.compute.manager [-] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Took 1.26 seconds to deallocate network for instance. [ 1005.646848] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1e734b28-52cc-4ed6-8f91-4dee8c3dda0d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.658071] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76ae9b8e-00ba-4da9-b783-3070ef322027 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.688263] env[62460]: DEBUG nova.compute.manager [req-0c0d2bfb-cc75-41e4-9d2f-c6c63274388c req-15402eb5-c670-4ea0-b014-8922bd713a9b service nova] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Detach interface failed, port_id=786ef177-843f-44d0-8920-2f4332e29154, reason: Instance 0d39a304-6a49-4d87-bfa2-4ba3b383578a could not be found. {{(pid=62460) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1005.695171] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7c85a888-2a2d-4bed-94e6-874a37ef74e2 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lock "896f8ff4-e45f-4403-a727-03ee25e58609" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.815s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.797220] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fae854a-4610-40a6-92de-2f73cfa0ca33 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.805555] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e204cc5f-5aa3-4cf1-891b-c8b5685dab1f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.841619] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73505bcb-c5bb-4075-96ef-e24eed04ad5f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.851410] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7e7e4dc-8db4-4489-8912-de0ba5fa81e2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.865531] env[62460]: DEBUG nova.compute.provider_tree [None req-b7594f06-69f8-4a0a-b7cc-5eecc483a3d6 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1006.143889] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b9b15fa3-0ba6-440c-9cb3-48dc854f853f tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.368633] env[62460]: DEBUG nova.scheduler.client.report [None req-b7594f06-69f8-4a0a-b7cc-5eecc483a3d6 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1006.639729] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Acquiring lock "c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.639986] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Lock "c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.874094] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b7594f06-69f8-4a0a-b7cc-5eecc483a3d6 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.722s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.877838] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.333s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.879227] env[62460]: INFO nova.compute.claims [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1006.901713] env[62460]: INFO nova.scheduler.client.report [None req-b7594f06-69f8-4a0a-b7cc-5eecc483a3d6 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Deleted allocations for instance 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea [ 1007.142959] env[62460]: DEBUG nova.compute.manager [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1007.412684] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b7594f06-69f8-4a0a-b7cc-5eecc483a3d6 tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.010s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.663037] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.983070] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dddfbb61-413f-47ac-9da6-54c5f03e64b3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.990884] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29670d4f-c542-4c98-97a0-6caa52512836 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.020020] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e8c5d72-763b-41a9-9368-66484bcf4b21 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.026824] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2569eaf7-e334-4663-b132-d5b8fea812dc {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.040472] env[62460]: DEBUG nova.compute.provider_tree [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1008.154473] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquiring lock "00597ea8-88ee-448d-b5b5-2b29dc8683e9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.154728] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lock "00597ea8-88ee-448d-b5b5-2b29dc8683e9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.238139] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c5dce603-3517-4c0b-af6b-211990e1a34a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "abd4501b-3ad8-4103-921d-5b80fc2f313f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.238610] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c5dce603-3517-4c0b-af6b-211990e1a34a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "abd4501b-3ad8-4103-921d-5b80fc2f313f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.238845] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c5dce603-3517-4c0b-af6b-211990e1a34a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "abd4501b-3ad8-4103-921d-5b80fc2f313f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.239672] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c5dce603-3517-4c0b-af6b-211990e1a34a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "abd4501b-3ad8-4103-921d-5b80fc2f313f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.239934] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c5dce603-3517-4c0b-af6b-211990e1a34a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "abd4501b-3ad8-4103-921d-5b80fc2f313f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.242176] env[62460]: INFO nova.compute.manager [None req-c5dce603-3517-4c0b-af6b-211990e1a34a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Terminating instance [ 1008.244853] env[62460]: DEBUG nova.compute.manager [None req-c5dce603-3517-4c0b-af6b-211990e1a34a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1008.245141] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-c5dce603-3517-4c0b-af6b-211990e1a34a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1008.245933] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38fe2634-5016-4147-9d5b-4e4b8cd4afa1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.253792] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5dce603-3517-4c0b-af6b-211990e1a34a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1008.254017] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-93fca514-e094-42db-bba1-9dd64d2be91f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.259576] env[62460]: DEBUG oslo_vmware.api [None req-c5dce603-3517-4c0b-af6b-211990e1a34a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Waiting for the task: (returnval){ [ 1008.259576] env[62460]: value = "task-1314034" [ 1008.259576] env[62460]: _type = "Task" [ 1008.259576] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.267929] env[62460]: DEBUG oslo_vmware.api [None req-c5dce603-3517-4c0b-af6b-211990e1a34a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1314034, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.544044] env[62460]: DEBUG nova.scheduler.client.report [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1008.657748] env[62460]: DEBUG nova.compute.manager [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1008.769936] env[62460]: DEBUG oslo_vmware.api [None req-c5dce603-3517-4c0b-af6b-211990e1a34a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1314034, 'name': PowerOffVM_Task, 'duration_secs': 0.216445} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.770297] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5dce603-3517-4c0b-af6b-211990e1a34a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1008.770513] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-c5dce603-3517-4c0b-af6b-211990e1a34a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1008.770778] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a2126bf7-f6f5-4fc3-8913-4f63033aae46 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.880746] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-c5dce603-3517-4c0b-af6b-211990e1a34a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1008.880986] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-c5dce603-3517-4c0b-af6b-211990e1a34a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Deleting contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1008.881199] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5dce603-3517-4c0b-af6b-211990e1a34a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Deleting the datastore file [datastore2] abd4501b-3ad8-4103-921d-5b80fc2f313f {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1008.881520] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-38ddcfa5-727a-4c2d-a0de-0feab96d4192 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.888947] env[62460]: DEBUG oslo_vmware.api [None req-c5dce603-3517-4c0b-af6b-211990e1a34a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Waiting for the task: (returnval){ [ 1008.888947] env[62460]: value = "task-1314036" [ 1008.888947] env[62460]: _type = "Task" [ 1008.888947] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.896691] env[62460]: DEBUG oslo_vmware.api [None req-c5dce603-3517-4c0b-af6b-211990e1a34a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1314036, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.048701] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.171s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.049323] env[62460]: DEBUG nova.compute.manager [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1009.052015] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b9b15fa3-0ba6-440c-9cb3-48dc854f853f tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.909s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.052281] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b9b15fa3-0ba6-440c-9cb3-48dc854f853f tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.054298] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.391s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.055728] env[62460]: INFO nova.compute.claims [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1009.073494] env[62460]: INFO nova.scheduler.client.report [None req-b9b15fa3-0ba6-440c-9cb3-48dc854f853f tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Deleted allocations for instance 0d39a304-6a49-4d87-bfa2-4ba3b383578a [ 1009.178433] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.398788] env[62460]: DEBUG oslo_vmware.api [None req-c5dce603-3517-4c0b-af6b-211990e1a34a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Task: {'id': task-1314036, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.15425} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.399062] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5dce603-3517-4c0b-af6b-211990e1a34a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1009.399264] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-c5dce603-3517-4c0b-af6b-211990e1a34a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Deleted contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1009.399457] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-c5dce603-3517-4c0b-af6b-211990e1a34a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1009.400198] env[62460]: INFO nova.compute.manager [None req-c5dce603-3517-4c0b-af6b-211990e1a34a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1009.400198] env[62460]: DEBUG oslo.service.loopingcall [None req-c5dce603-3517-4c0b-af6b-211990e1a34a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1009.400198] env[62460]: DEBUG nova.compute.manager [-] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1009.400341] env[62460]: DEBUG nova.network.neutron [-] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1009.560291] env[62460]: DEBUG nova.compute.utils [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1009.564071] env[62460]: DEBUG nova.compute.manager [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1009.564071] env[62460]: DEBUG nova.network.neutron [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1009.582369] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b9b15fa3-0ba6-440c-9cb3-48dc854f853f tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "0d39a304-6a49-4d87-bfa2-4ba3b383578a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.543s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.603163] env[62460]: DEBUG nova.policy [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e531b7c3e7544e53b126d1336ac44c99', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bfa5a65b0f614e769de5b3aa77bf869e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 1009.637367] env[62460]: DEBUG nova.compute.manager [req-20b9e439-ebd7-4853-87bd-8324f9db7c37 req-ade8b309-cc74-402c-a6be-a360ad273c16 service nova] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Received event network-vif-deleted-546b0347-51db-4a4a-99e9-4f3db96a36bd {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1009.637367] env[62460]: INFO nova.compute.manager [req-20b9e439-ebd7-4853-87bd-8324f9db7c37 req-ade8b309-cc74-402c-a6be-a360ad273c16 service nova] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Neutron deleted interface 546b0347-51db-4a4a-99e9-4f3db96a36bd; detaching it from the instance and deleting it from the info cache [ 1009.637550] env[62460]: DEBUG nova.network.neutron [req-20b9e439-ebd7-4853-87bd-8324f9db7c37 req-ade8b309-cc74-402c-a6be-a360ad273c16 service nova] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.831285] env[62460]: DEBUG nova.network.neutron [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Successfully created port: f516ed8d-249e-4717-bb01-d98f9e81cef5 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1010.064755] env[62460]: DEBUG nova.compute.manager [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1010.116070] env[62460]: DEBUG nova.network.neutron [-] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1010.141039] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f461e970-c40a-40db-9c38-732471a2e409 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.149991] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02ed8b0a-8988-4fca-b6a7-13fa7ece16dd {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.161724] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-914de762-9af1-4fca-a36b-f2f68a4ff390 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.169251] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0974fd09-4514-450c-8f4b-0be5d4b2bcae {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.177057] env[62460]: DEBUG nova.compute.manager [req-20b9e439-ebd7-4853-87bd-8324f9db7c37 req-ade8b309-cc74-402c-a6be-a360ad273c16 service nova] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Detach interface failed, port_id=546b0347-51db-4a4a-99e9-4f3db96a36bd, reason: Instance abd4501b-3ad8-4103-921d-5b80fc2f313f could not be found. {{(pid=62460) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1010.201439] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29ad1976-498a-4fe9-9b88-f5fdf13474ab {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.208738] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42e77b78-9d44-4fbf-908e-9e2f5418d044 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.223618] env[62460]: DEBUG nova.compute.provider_tree [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1010.620246] env[62460]: INFO nova.compute.manager [-] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Took 1.22 seconds to deallocate network for instance. [ 1010.727750] env[62460]: DEBUG nova.scheduler.client.report [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1011.077550] env[62460]: DEBUG nova.compute.manager [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1011.102595] env[62460]: DEBUG nova.virt.hardware [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1011.102878] env[62460]: DEBUG nova.virt.hardware [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1011.103075] env[62460]: DEBUG nova.virt.hardware [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1011.103279] env[62460]: DEBUG nova.virt.hardware [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1011.103436] env[62460]: DEBUG nova.virt.hardware [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1011.103594] env[62460]: DEBUG nova.virt.hardware [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1011.103807] env[62460]: DEBUG nova.virt.hardware [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1011.103981] env[62460]: DEBUG nova.virt.hardware [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1011.104173] env[62460]: DEBUG nova.virt.hardware [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1011.104347] env[62460]: DEBUG nova.virt.hardware [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1011.104538] env[62460]: DEBUG nova.virt.hardware [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1011.105429] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f740556-9d85-439f-a2ae-52e04ead7568 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.113420] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5d07065-5bab-4e13-bed6-83a11c476615 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.127216] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c5dce603-3517-4c0b-af6b-211990e1a34a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.215666] env[62460]: DEBUG nova.compute.manager [req-4fd63c2b-c9c0-4820-8290-6568e9663c20 req-1a7065be-1941-4f3b-b9ef-f0d6d1a0c026 service nova] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Received event network-vif-plugged-f516ed8d-249e-4717-bb01-d98f9e81cef5 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1011.215911] env[62460]: DEBUG oslo_concurrency.lockutils [req-4fd63c2b-c9c0-4820-8290-6568e9663c20 req-1a7065be-1941-4f3b-b9ef-f0d6d1a0c026 service nova] Acquiring lock "37e76fbf-fc84-4f63-ba7f-3515ccb376c1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.216143] env[62460]: DEBUG oslo_concurrency.lockutils [req-4fd63c2b-c9c0-4820-8290-6568e9663c20 req-1a7065be-1941-4f3b-b9ef-f0d6d1a0c026 service nova] Lock "37e76fbf-fc84-4f63-ba7f-3515ccb376c1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.216352] env[62460]: DEBUG oslo_concurrency.lockutils [req-4fd63c2b-c9c0-4820-8290-6568e9663c20 req-1a7065be-1941-4f3b-b9ef-f0d6d1a0c026 service nova] Lock "37e76fbf-fc84-4f63-ba7f-3515ccb376c1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.216547] env[62460]: DEBUG nova.compute.manager [req-4fd63c2b-c9c0-4820-8290-6568e9663c20 req-1a7065be-1941-4f3b-b9ef-f0d6d1a0c026 service nova] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] No waiting events found dispatching network-vif-plugged-f516ed8d-249e-4717-bb01-d98f9e81cef5 {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1011.216732] env[62460]: WARNING nova.compute.manager [req-4fd63c2b-c9c0-4820-8290-6568e9663c20 req-1a7065be-1941-4f3b-b9ef-f0d6d1a0c026 service nova] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Received unexpected event network-vif-plugged-f516ed8d-249e-4717-bb01-d98f9e81cef5 for instance with vm_state building and task_state spawning. [ 1011.231553] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.177s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.232160] env[62460]: DEBUG nova.compute.manager [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1011.234575] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.056s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.238028] env[62460]: INFO nova.compute.claims [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1011.305931] env[62460]: DEBUG nova.network.neutron [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Successfully updated port: f516ed8d-249e-4717-bb01-d98f9e81cef5 {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1011.740177] env[62460]: DEBUG nova.compute.utils [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1011.744677] env[62460]: DEBUG nova.compute.manager [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1011.744858] env[62460]: DEBUG nova.network.neutron [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1011.783400] env[62460]: DEBUG nova.policy [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd2b0dac2c7fa4b60b547a3f081e27980', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4f13597d128c44e19b7f92f5ba1ba217', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 1011.807828] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "refresh_cache-37e76fbf-fc84-4f63-ba7f-3515ccb376c1" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1011.808113] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquired lock "refresh_cache-37e76fbf-fc84-4f63-ba7f-3515ccb376c1" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.808192] env[62460]: DEBUG nova.network.neutron [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1012.029836] env[62460]: DEBUG nova.network.neutron [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Successfully created port: d15647d8-f541-4169-bea9-0937db3301b1 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1012.245395] env[62460]: DEBUG nova.compute.manager [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1012.346822] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de93d1fa-6007-48f5-8ad2-2c0ae1a4dfb5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.352048] env[62460]: DEBUG nova.network.neutron [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1012.359716] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9a60931-fc1e-481f-b931-08e69d09cbde {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.392023] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-448af79a-5530-4f0c-9c82-916653164fcc {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.398855] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17861cd5-1e38-42f5-aaa5-0d68d68c1bc4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.412263] env[62460]: DEBUG nova.compute.provider_tree [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1012.508662] env[62460]: DEBUG nova.network.neutron [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Updating instance_info_cache with network_info: [{"id": "f516ed8d-249e-4717-bb01-d98f9e81cef5", "address": "fa:16:3e:0e:69:8a", "network": {"id": "3c49720c-94c9-4e2f-bbb7-4266cc97faee", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1650281790-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfa5a65b0f614e769de5b3aa77bf869e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf516ed8d-24", "ovs_interfaceid": "f516ed8d-249e-4717-bb01-d98f9e81cef5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.915749] env[62460]: DEBUG nova.scheduler.client.report [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1013.011027] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Releasing lock "refresh_cache-37e76fbf-fc84-4f63-ba7f-3515ccb376c1" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.011402] env[62460]: DEBUG nova.compute.manager [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Instance network_info: |[{"id": "f516ed8d-249e-4717-bb01-d98f9e81cef5", "address": "fa:16:3e:0e:69:8a", "network": {"id": "3c49720c-94c9-4e2f-bbb7-4266cc97faee", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1650281790-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfa5a65b0f614e769de5b3aa77bf869e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf516ed8d-24", "ovs_interfaceid": "f516ed8d-249e-4717-bb01-d98f9e81cef5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1013.011844] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0e:69:8a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '04ccbc7a-cf8d-4ea2-8411-291a1e27df7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f516ed8d-249e-4717-bb01-d98f9e81cef5', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1013.019993] env[62460]: DEBUG oslo.service.loopingcall [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1013.020226] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1013.020456] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ee3c035d-be36-477e-9fac-56f8b9654d1c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.041082] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1013.041082] env[62460]: value = "task-1314037" [ 1013.041082] env[62460]: _type = "Task" [ 1013.041082] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.048785] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1314037, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.242600] env[62460]: DEBUG nova.compute.manager [req-d3b6ee24-48a4-4e61-badf-c21d9c61711e req-8bb4b85d-8a06-4440-b0cf-b309c62def65 service nova] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Received event network-changed-f516ed8d-249e-4717-bb01-d98f9e81cef5 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1013.242873] env[62460]: DEBUG nova.compute.manager [req-d3b6ee24-48a4-4e61-badf-c21d9c61711e req-8bb4b85d-8a06-4440-b0cf-b309c62def65 service nova] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Refreshing instance network info cache due to event network-changed-f516ed8d-249e-4717-bb01-d98f9e81cef5. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1013.243185] env[62460]: DEBUG oslo_concurrency.lockutils [req-d3b6ee24-48a4-4e61-badf-c21d9c61711e req-8bb4b85d-8a06-4440-b0cf-b309c62def65 service nova] Acquiring lock "refresh_cache-37e76fbf-fc84-4f63-ba7f-3515ccb376c1" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1013.243468] env[62460]: DEBUG oslo_concurrency.lockutils [req-d3b6ee24-48a4-4e61-badf-c21d9c61711e req-8bb4b85d-8a06-4440-b0cf-b309c62def65 service nova] Acquired lock "refresh_cache-37e76fbf-fc84-4f63-ba7f-3515ccb376c1" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.243693] env[62460]: DEBUG nova.network.neutron [req-d3b6ee24-48a4-4e61-badf-c21d9c61711e req-8bb4b85d-8a06-4440-b0cf-b309c62def65 service nova] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Refreshing network info cache for port f516ed8d-249e-4717-bb01-d98f9e81cef5 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1013.257277] env[62460]: DEBUG nova.compute.manager [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1013.283361] env[62460]: DEBUG nova.virt.hardware [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1013.283619] env[62460]: DEBUG nova.virt.hardware [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1013.283795] env[62460]: DEBUG nova.virt.hardware [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1013.283979] env[62460]: DEBUG nova.virt.hardware [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1013.284151] env[62460]: DEBUG nova.virt.hardware [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1013.284306] env[62460]: DEBUG nova.virt.hardware [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1013.284520] env[62460]: DEBUG nova.virt.hardware [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1013.284694] env[62460]: DEBUG nova.virt.hardware [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1013.284866] env[62460]: DEBUG nova.virt.hardware [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1013.285109] env[62460]: DEBUG nova.virt.hardware [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1013.285360] env[62460]: DEBUG nova.virt.hardware [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1013.286600] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f444b75-09e3-409d-941d-5ac255cdbde8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.296379] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ab21183-6b0d-401e-80a9-8450da84a7c6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.420252] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.185s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.420888] env[62460]: DEBUG nova.compute.manager [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1013.423813] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c5dce603-3517-4c0b-af6b-211990e1a34a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.297s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.424065] env[62460]: DEBUG nova.objects.instance [None req-c5dce603-3517-4c0b-af6b-211990e1a34a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lazy-loading 'resources' on Instance uuid abd4501b-3ad8-4103-921d-5b80fc2f313f {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1013.484041] env[62460]: DEBUG nova.network.neutron [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Successfully updated port: d15647d8-f541-4169-bea9-0937db3301b1 {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1013.551640] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1314037, 'name': CreateVM_Task, 'duration_secs': 0.289642} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.551640] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1013.552710] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1013.552710] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.552710] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1013.552938] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8d49a857-8cbe-48ee-867b-7d19b98a8ca8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.557090] env[62460]: DEBUG oslo_vmware.api [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 1013.557090] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52002cbb-d73f-159a-66cf-de00add7e1c3" [ 1013.557090] env[62460]: _type = "Task" [ 1013.557090] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.564717] env[62460]: DEBUG oslo_vmware.api [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52002cbb-d73f-159a-66cf-de00add7e1c3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.927653] env[62460]: DEBUG nova.compute.utils [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1013.934871] env[62460]: DEBUG nova.compute.manager [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1013.934871] env[62460]: DEBUG nova.network.neutron [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1013.988092] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Acquiring lock "refresh_cache-c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1013.988092] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Acquired lock "refresh_cache-c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.988092] env[62460]: DEBUG nova.network.neutron [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1013.994372] env[62460]: DEBUG nova.policy [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '947371ed3ed94ee5b2a900d47444791c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f3d2d1c48ec14121a2e8c9b3f800a949', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 1013.994372] env[62460]: DEBUG nova.network.neutron [req-d3b6ee24-48a4-4e61-badf-c21d9c61711e req-8bb4b85d-8a06-4440-b0cf-b309c62def65 service nova] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Updated VIF entry in instance network info cache for port f516ed8d-249e-4717-bb01-d98f9e81cef5. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1013.994372] env[62460]: DEBUG nova.network.neutron [req-d3b6ee24-48a4-4e61-badf-c21d9c61711e req-8bb4b85d-8a06-4440-b0cf-b309c62def65 service nova] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Updating instance_info_cache with network_info: [{"id": "f516ed8d-249e-4717-bb01-d98f9e81cef5", "address": "fa:16:3e:0e:69:8a", "network": {"id": "3c49720c-94c9-4e2f-bbb7-4266cc97faee", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1650281790-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfa5a65b0f614e769de5b3aa77bf869e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf516ed8d-24", "ovs_interfaceid": "f516ed8d-249e-4717-bb01-d98f9e81cef5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1014.032306] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab597994-4706-4403-bfbb-efa1c37fa30c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.039971] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63c48bda-aab3-4146-8a49-a8cd14d4c80f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.073961] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62241bcb-8b3c-4cf5-b9fd-b34dc70c3d0a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.082203] env[62460]: DEBUG oslo_vmware.api [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52002cbb-d73f-159a-66cf-de00add7e1c3, 'name': SearchDatastore_Task, 'duration_secs': 0.010948} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.083962] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1014.084224] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1014.084465] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.084627] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.084821] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1014.085116] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3760af88-b5ba-46fa-8d95-cd0738635008 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.087567] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e39320e0-67fa-485a-be8f-15a57a099288 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.100449] env[62460]: DEBUG nova.compute.provider_tree [None req-c5dce603-3517-4c0b-af6b-211990e1a34a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1014.102540] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1014.102727] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1014.103608] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-38a9c415-f45c-46ca-816f-501ca619bd7a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.108521] env[62460]: DEBUG oslo_vmware.api [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 1014.108521] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]522e97b4-04d0-c98b-9d32-5e30bbcdf9ff" [ 1014.108521] env[62460]: _type = "Task" [ 1014.108521] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.116423] env[62460]: DEBUG oslo_vmware.api [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]522e97b4-04d0-c98b-9d32-5e30bbcdf9ff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.257851] env[62460]: DEBUG nova.network.neutron [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Successfully created port: c591ec72-086f-44a0-8cf2-3e6719700505 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1014.435669] env[62460]: DEBUG nova.compute.manager [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1014.498476] env[62460]: DEBUG oslo_concurrency.lockutils [req-d3b6ee24-48a4-4e61-badf-c21d9c61711e req-8bb4b85d-8a06-4440-b0cf-b309c62def65 service nova] Releasing lock "refresh_cache-37e76fbf-fc84-4f63-ba7f-3515ccb376c1" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1014.525660] env[62460]: DEBUG nova.network.neutron [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1014.605025] env[62460]: DEBUG nova.scheduler.client.report [None req-c5dce603-3517-4c0b-af6b-211990e1a34a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1014.620472] env[62460]: DEBUG oslo_vmware.api [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]522e97b4-04d0-c98b-9d32-5e30bbcdf9ff, 'name': SearchDatastore_Task, 'duration_secs': 0.008533} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.621842] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f9b708f-9cf5-47d7-990b-2e0399c8b536 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.627228] env[62460]: DEBUG oslo_vmware.api [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 1014.627228] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]523dcc11-f05e-8ea9-f053-b3c512be0977" [ 1014.627228] env[62460]: _type = "Task" [ 1014.627228] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.635218] env[62460]: DEBUG oslo_vmware.api [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]523dcc11-f05e-8ea9-f053-b3c512be0977, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.653372] env[62460]: DEBUG nova.network.neutron [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Updating instance_info_cache with network_info: [{"id": "d15647d8-f541-4169-bea9-0937db3301b1", "address": "fa:16:3e:28:b4:d2", "network": {"id": "3aa5e026-c4ee-4641-831b-a7e31b69adc7", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-702747461-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f13597d128c44e19b7f92f5ba1ba217", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b91b49a8-b849-4d0c-97f7-74fdcd88ae03", "external-id": "nsx-vlan-transportzone-406", "segmentation_id": 406, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd15647d8-f5", "ovs_interfaceid": "d15647d8-f541-4169-bea9-0937db3301b1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.110046] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c5dce603-3517-4c0b-af6b-211990e1a34a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.686s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.127511] env[62460]: INFO nova.scheduler.client.report [None req-c5dce603-3517-4c0b-af6b-211990e1a34a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Deleted allocations for instance abd4501b-3ad8-4103-921d-5b80fc2f313f [ 1015.140581] env[62460]: DEBUG oslo_vmware.api [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]523dcc11-f05e-8ea9-f053-b3c512be0977, 'name': SearchDatastore_Task, 'duration_secs': 0.008859} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.140975] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.141317] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] 37e76fbf-fc84-4f63-ba7f-3515ccb376c1/37e76fbf-fc84-4f63-ba7f-3515ccb376c1.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1015.141625] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c9d3a314-9614-4e2b-ba33-a0e25ef07095 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.149276] env[62460]: DEBUG oslo_vmware.api [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 1015.149276] env[62460]: value = "task-1314038" [ 1015.149276] env[62460]: _type = "Task" [ 1015.149276] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.157057] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Releasing lock "refresh_cache-c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.157438] env[62460]: DEBUG nova.compute.manager [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Instance network_info: |[{"id": "d15647d8-f541-4169-bea9-0937db3301b1", "address": "fa:16:3e:28:b4:d2", "network": {"id": "3aa5e026-c4ee-4641-831b-a7e31b69adc7", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-702747461-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f13597d128c44e19b7f92f5ba1ba217", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b91b49a8-b849-4d0c-97f7-74fdcd88ae03", "external-id": "nsx-vlan-transportzone-406", "segmentation_id": 406, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd15647d8-f5", "ovs_interfaceid": "d15647d8-f541-4169-bea9-0937db3301b1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1015.157775] env[62460]: DEBUG oslo_vmware.api [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1314038, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.158211] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:28:b4:d2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b91b49a8-b849-4d0c-97f7-74fdcd88ae03', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd15647d8-f541-4169-bea9-0937db3301b1', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1015.166095] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Creating folder: Project (4f13597d128c44e19b7f92f5ba1ba217). Parent ref: group-v281134. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1015.166411] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-60a5c43f-727d-4872-9c99-9436978220a5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.177330] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Created folder: Project (4f13597d128c44e19b7f92f5ba1ba217) in parent group-v281134. [ 1015.177565] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Creating folder: Instances. Parent ref: group-v281266. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1015.177830] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3f91219b-e7e2-44de-8f54-c90421306bdd {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.188990] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Created folder: Instances in parent group-v281266. [ 1015.189270] env[62460]: DEBUG oslo.service.loopingcall [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1015.189493] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1015.189732] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3b98247e-f5df-4b52-aa41-172602fb26d6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.208356] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1015.208356] env[62460]: value = "task-1314041" [ 1015.208356] env[62460]: _type = "Task" [ 1015.208356] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.220124] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1314041, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.272988] env[62460]: DEBUG nova.compute.manager [req-424045f2-806f-4847-b2a3-ba9d75354074 req-7f4b50b2-f336-45b1-bfbc-21671430a971 service nova] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Received event network-vif-plugged-d15647d8-f541-4169-bea9-0937db3301b1 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1015.273184] env[62460]: DEBUG oslo_concurrency.lockutils [req-424045f2-806f-4847-b2a3-ba9d75354074 req-7f4b50b2-f336-45b1-bfbc-21671430a971 service nova] Acquiring lock "c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.273427] env[62460]: DEBUG oslo_concurrency.lockutils [req-424045f2-806f-4847-b2a3-ba9d75354074 req-7f4b50b2-f336-45b1-bfbc-21671430a971 service nova] Lock "c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.273621] env[62460]: DEBUG oslo_concurrency.lockutils [req-424045f2-806f-4847-b2a3-ba9d75354074 req-7f4b50b2-f336-45b1-bfbc-21671430a971 service nova] Lock "c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.273834] env[62460]: DEBUG nova.compute.manager [req-424045f2-806f-4847-b2a3-ba9d75354074 req-7f4b50b2-f336-45b1-bfbc-21671430a971 service nova] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] No waiting events found dispatching network-vif-plugged-d15647d8-f541-4169-bea9-0937db3301b1 {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1015.274039] env[62460]: WARNING nova.compute.manager [req-424045f2-806f-4847-b2a3-ba9d75354074 req-7f4b50b2-f336-45b1-bfbc-21671430a971 service nova] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Received unexpected event network-vif-plugged-d15647d8-f541-4169-bea9-0937db3301b1 for instance with vm_state building and task_state spawning. [ 1015.274225] env[62460]: DEBUG nova.compute.manager [req-424045f2-806f-4847-b2a3-ba9d75354074 req-7f4b50b2-f336-45b1-bfbc-21671430a971 service nova] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Received event network-changed-d15647d8-f541-4169-bea9-0937db3301b1 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1015.274415] env[62460]: DEBUG nova.compute.manager [req-424045f2-806f-4847-b2a3-ba9d75354074 req-7f4b50b2-f336-45b1-bfbc-21671430a971 service nova] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Refreshing instance network info cache due to event network-changed-d15647d8-f541-4169-bea9-0937db3301b1. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1015.274629] env[62460]: DEBUG oslo_concurrency.lockutils [req-424045f2-806f-4847-b2a3-ba9d75354074 req-7f4b50b2-f336-45b1-bfbc-21671430a971 service nova] Acquiring lock "refresh_cache-c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.274783] env[62460]: DEBUG oslo_concurrency.lockutils [req-424045f2-806f-4847-b2a3-ba9d75354074 req-7f4b50b2-f336-45b1-bfbc-21671430a971 service nova] Acquired lock "refresh_cache-c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.274982] env[62460]: DEBUG nova.network.neutron [req-424045f2-806f-4847-b2a3-ba9d75354074 req-7f4b50b2-f336-45b1-bfbc-21671430a971 service nova] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Refreshing network info cache for port d15647d8-f541-4169-bea9-0937db3301b1 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1015.450188] env[62460]: DEBUG nova.compute.manager [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1015.477550] env[62460]: DEBUG nova.virt.hardware [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1015.477800] env[62460]: DEBUG nova.virt.hardware [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1015.478093] env[62460]: DEBUG nova.virt.hardware [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1015.478165] env[62460]: DEBUG nova.virt.hardware [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1015.478322] env[62460]: DEBUG nova.virt.hardware [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1015.478490] env[62460]: DEBUG nova.virt.hardware [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1015.478741] env[62460]: DEBUG nova.virt.hardware [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1015.478916] env[62460]: DEBUG nova.virt.hardware [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1015.479103] env[62460]: DEBUG nova.virt.hardware [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1015.479284] env[62460]: DEBUG nova.virt.hardware [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1015.479466] env[62460]: DEBUG nova.virt.hardware [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1015.480416] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a87a980b-1486-4b61-935a-2b08791bbe0c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.491203] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-541108fa-8be7-4d23-8069-8a7c66cb7eb7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.639851] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c5dce603-3517-4c0b-af6b-211990e1a34a tempest-AttachInterfacesTestJSON-767722393 tempest-AttachInterfacesTestJSON-767722393-project-member] Lock "abd4501b-3ad8-4103-921d-5b80fc2f313f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.400s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.659390] env[62460]: DEBUG oslo_vmware.api [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1314038, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.479659} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.660293] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] 37e76fbf-fc84-4f63-ba7f-3515ccb376c1/37e76fbf-fc84-4f63-ba7f-3515ccb376c1.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1015.660427] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1015.660679] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a91b9f0e-209e-4bdf-8b3f-c86d6543087f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.667219] env[62460]: DEBUG oslo_vmware.api [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 1015.667219] env[62460]: value = "task-1314042" [ 1015.667219] env[62460]: _type = "Task" [ 1015.667219] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.675626] env[62460]: DEBUG oslo_vmware.api [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1314042, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.702367] env[62460]: DEBUG nova.compute.manager [req-1251561c-d0df-4c56-9b8f-3966e2d022d3 req-f01d0afa-c379-4033-b739-76a311bd22a1 service nova] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Received event network-vif-plugged-c591ec72-086f-44a0-8cf2-3e6719700505 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1015.702588] env[62460]: DEBUG oslo_concurrency.lockutils [req-1251561c-d0df-4c56-9b8f-3966e2d022d3 req-f01d0afa-c379-4033-b739-76a311bd22a1 service nova] Acquiring lock "00597ea8-88ee-448d-b5b5-2b29dc8683e9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.702802] env[62460]: DEBUG oslo_concurrency.lockutils [req-1251561c-d0df-4c56-9b8f-3966e2d022d3 req-f01d0afa-c379-4033-b739-76a311bd22a1 service nova] Lock "00597ea8-88ee-448d-b5b5-2b29dc8683e9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.702977] env[62460]: DEBUG oslo_concurrency.lockutils [req-1251561c-d0df-4c56-9b8f-3966e2d022d3 req-f01d0afa-c379-4033-b739-76a311bd22a1 service nova] Lock "00597ea8-88ee-448d-b5b5-2b29dc8683e9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.703170] env[62460]: DEBUG nova.compute.manager [req-1251561c-d0df-4c56-9b8f-3966e2d022d3 req-f01d0afa-c379-4033-b739-76a311bd22a1 service nova] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] No waiting events found dispatching network-vif-plugged-c591ec72-086f-44a0-8cf2-3e6719700505 {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1015.703346] env[62460]: WARNING nova.compute.manager [req-1251561c-d0df-4c56-9b8f-3966e2d022d3 req-f01d0afa-c379-4033-b739-76a311bd22a1 service nova] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Received unexpected event network-vif-plugged-c591ec72-086f-44a0-8cf2-3e6719700505 for instance with vm_state building and task_state spawning. [ 1015.718062] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1314041, 'name': CreateVM_Task, 'duration_secs': 0.468018} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.718238] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1015.718898] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.719085] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.719420] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1015.719683] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-63483ace-904d-4008-ab05-bb019724c82d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.724994] env[62460]: DEBUG oslo_vmware.api [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Waiting for the task: (returnval){ [ 1015.724994] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52de4186-c222-b038-49b9-e3e05af7d04d" [ 1015.724994] env[62460]: _type = "Task" [ 1015.724994] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.733219] env[62460]: DEBUG oslo_vmware.api [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52de4186-c222-b038-49b9-e3e05af7d04d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.786805] env[62460]: DEBUG nova.network.neutron [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Successfully updated port: c591ec72-086f-44a0-8cf2-3e6719700505 {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1016.066318] env[62460]: DEBUG nova.network.neutron [req-424045f2-806f-4847-b2a3-ba9d75354074 req-7f4b50b2-f336-45b1-bfbc-21671430a971 service nova] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Updated VIF entry in instance network info cache for port d15647d8-f541-4169-bea9-0937db3301b1. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1016.066722] env[62460]: DEBUG nova.network.neutron [req-424045f2-806f-4847-b2a3-ba9d75354074 req-7f4b50b2-f336-45b1-bfbc-21671430a971 service nova] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Updating instance_info_cache with network_info: [{"id": "d15647d8-f541-4169-bea9-0937db3301b1", "address": "fa:16:3e:28:b4:d2", "network": {"id": "3aa5e026-c4ee-4641-831b-a7e31b69adc7", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-702747461-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f13597d128c44e19b7f92f5ba1ba217", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b91b49a8-b849-4d0c-97f7-74fdcd88ae03", "external-id": "nsx-vlan-transportzone-406", "segmentation_id": 406, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd15647d8-f5", "ovs_interfaceid": "d15647d8-f541-4169-bea9-0937db3301b1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1016.177036] env[62460]: DEBUG oslo_vmware.api [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1314042, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061718} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.177330] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1016.178092] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f342a922-6efc-49d5-b7ee-3b93b13ea679 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.199347] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Reconfiguring VM instance instance-00000065 to attach disk [datastore2] 37e76fbf-fc84-4f63-ba7f-3515ccb376c1/37e76fbf-fc84-4f63-ba7f-3515ccb376c1.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1016.199620] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3ad7c912-2049-4fa8-ad96-0ef0e84a8143 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.218453] env[62460]: DEBUG oslo_vmware.api [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 1016.218453] env[62460]: value = "task-1314043" [ 1016.218453] env[62460]: _type = "Task" [ 1016.218453] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.228184] env[62460]: DEBUG oslo_vmware.api [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1314043, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.236575] env[62460]: DEBUG oslo_vmware.api [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52de4186-c222-b038-49b9-e3e05af7d04d, 'name': SearchDatastore_Task, 'duration_secs': 0.008758} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.237186] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.237591] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1016.238132] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1016.238478] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.238844] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1016.239339] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-12440c4b-3ec3-4da9-a18d-e94b2bbe1d0c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.248228] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1016.248566] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1016.249442] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a9214606-f3ec-4d6a-a2b4-bf5116aaa6cb {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.255170] env[62460]: DEBUG oslo_vmware.api [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Waiting for the task: (returnval){ [ 1016.255170] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52db2a1d-1b66-41fa-a056-b228c30480b3" [ 1016.255170] env[62460]: _type = "Task" [ 1016.255170] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.263314] env[62460]: DEBUG oslo_vmware.api [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52db2a1d-1b66-41fa-a056-b228c30480b3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.289551] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquiring lock "refresh_cache-00597ea8-88ee-448d-b5b5-2b29dc8683e9" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1016.289850] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquired lock "refresh_cache-00597ea8-88ee-448d-b5b5-2b29dc8683e9" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.290064] env[62460]: DEBUG nova.network.neutron [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1016.569593] env[62460]: DEBUG oslo_concurrency.lockutils [req-424045f2-806f-4847-b2a3-ba9d75354074 req-7f4b50b2-f336-45b1-bfbc-21671430a971 service nova] Releasing lock "refresh_cache-c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.732656] env[62460]: DEBUG oslo_vmware.api [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1314043, 'name': ReconfigVM_Task, 'duration_secs': 0.267057} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.734021] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Reconfigured VM instance instance-00000065 to attach disk [datastore2] 37e76fbf-fc84-4f63-ba7f-3515ccb376c1/37e76fbf-fc84-4f63-ba7f-3515ccb376c1.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1016.734515] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9b82a091-ed8b-4999-8eb8-81335a3b7c25 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.744363] env[62460]: DEBUG oslo_vmware.api [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 1016.744363] env[62460]: value = "task-1314044" [ 1016.744363] env[62460]: _type = "Task" [ 1016.744363] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.756074] env[62460]: DEBUG oslo_vmware.api [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1314044, 'name': Rename_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.768155] env[62460]: DEBUG oslo_vmware.api [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52db2a1d-1b66-41fa-a056-b228c30480b3, 'name': SearchDatastore_Task, 'duration_secs': 0.008294} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.768155] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b7cc841c-f963-4590-a0cc-b15f91431dd5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.773341] env[62460]: DEBUG oslo_vmware.api [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Waiting for the task: (returnval){ [ 1016.773341] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]526378d4-257e-8c83-51d5-c58a5e9184be" [ 1016.773341] env[62460]: _type = "Task" [ 1016.773341] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.780305] env[62460]: DEBUG oslo_vmware.api [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]526378d4-257e-8c83-51d5-c58a5e9184be, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.820499] env[62460]: DEBUG nova.network.neutron [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1016.957228] env[62460]: DEBUG nova.network.neutron [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Updating instance_info_cache with network_info: [{"id": "c591ec72-086f-44a0-8cf2-3e6719700505", "address": "fa:16:3e:30:d9:88", "network": {"id": "b5cb4ee0-4562-45be-a25d-51c3a5932efd", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-823340443-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3d2d1c48ec14121a2e8c9b3f800a949", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55c757ac-f8b2-466d-b634-07dbd100b312", "external-id": "nsx-vlan-transportzone-159", "segmentation_id": 159, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc591ec72-08", "ovs_interfaceid": "c591ec72-086f-44a0-8cf2-3e6719700505", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.257068] env[62460]: DEBUG oslo_vmware.api [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1314044, 'name': Rename_Task, 'duration_secs': 0.128271} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.257388] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1017.257652] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-09a2e89a-b4c9-4cfb-9196-83c14d1a4915 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.265195] env[62460]: DEBUG oslo_vmware.api [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 1017.265195] env[62460]: value = "task-1314045" [ 1017.265195] env[62460]: _type = "Task" [ 1017.265195] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.274888] env[62460]: DEBUG oslo_vmware.api [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1314045, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.284125] env[62460]: DEBUG oslo_vmware.api [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]526378d4-257e-8c83-51d5-c58a5e9184be, 'name': SearchDatastore_Task, 'duration_secs': 0.008588} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.284380] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1017.284652] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac/c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1017.284907] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-456e3695-d109-453f-886d-39928738867b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.294096] env[62460]: DEBUG oslo_vmware.api [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Waiting for the task: (returnval){ [ 1017.294096] env[62460]: value = "task-1314046" [ 1017.294096] env[62460]: _type = "Task" [ 1017.294096] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.303414] env[62460]: DEBUG oslo_vmware.api [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314046, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.463020] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Releasing lock "refresh_cache-00597ea8-88ee-448d-b5b5-2b29dc8683e9" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1017.463020] env[62460]: DEBUG nova.compute.manager [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Instance network_info: |[{"id": "c591ec72-086f-44a0-8cf2-3e6719700505", "address": "fa:16:3e:30:d9:88", "network": {"id": "b5cb4ee0-4562-45be-a25d-51c3a5932efd", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-823340443-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3d2d1c48ec14121a2e8c9b3f800a949", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55c757ac-f8b2-466d-b634-07dbd100b312", "external-id": "nsx-vlan-transportzone-159", "segmentation_id": 159, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc591ec72-08", "ovs_interfaceid": "c591ec72-086f-44a0-8cf2-3e6719700505", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1017.463020] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:30:d9:88', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '55c757ac-f8b2-466d-b634-07dbd100b312', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c591ec72-086f-44a0-8cf2-3e6719700505', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1017.469861] env[62460]: DEBUG oslo.service.loopingcall [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1017.470341] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1017.470737] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3c9494d5-7a84-45e9-a8aa-0268d0c05cef {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.493407] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1017.493407] env[62460]: value = "task-1314047" [ 1017.493407] env[62460]: _type = "Task" [ 1017.493407] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.502541] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1314047, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.745806] env[62460]: DEBUG nova.compute.manager [req-67b9f0b7-de5e-4404-9234-00279b98ceeb req-8af8a5d1-c69a-4ed0-a1be-98b080d606df service nova] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Received event network-changed-c591ec72-086f-44a0-8cf2-3e6719700505 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1017.746036] env[62460]: DEBUG nova.compute.manager [req-67b9f0b7-de5e-4404-9234-00279b98ceeb req-8af8a5d1-c69a-4ed0-a1be-98b080d606df service nova] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Refreshing instance network info cache due to event network-changed-c591ec72-086f-44a0-8cf2-3e6719700505. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1017.746319] env[62460]: DEBUG oslo_concurrency.lockutils [req-67b9f0b7-de5e-4404-9234-00279b98ceeb req-8af8a5d1-c69a-4ed0-a1be-98b080d606df service nova] Acquiring lock "refresh_cache-00597ea8-88ee-448d-b5b5-2b29dc8683e9" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1017.746469] env[62460]: DEBUG oslo_concurrency.lockutils [req-67b9f0b7-de5e-4404-9234-00279b98ceeb req-8af8a5d1-c69a-4ed0-a1be-98b080d606df service nova] Acquired lock "refresh_cache-00597ea8-88ee-448d-b5b5-2b29dc8683e9" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.746675] env[62460]: DEBUG nova.network.neutron [req-67b9f0b7-de5e-4404-9234-00279b98ceeb req-8af8a5d1-c69a-4ed0-a1be-98b080d606df service nova] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Refreshing network info cache for port c591ec72-086f-44a0-8cf2-3e6719700505 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1017.776500] env[62460]: DEBUG oslo_vmware.api [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1314045, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.805117] env[62460]: DEBUG oslo_vmware.api [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314046, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.462559} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.805508] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac/c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1017.805768] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1017.806130] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f8105373-cc87-4e80-9079-836b04422d5a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.813810] env[62460]: DEBUG oslo_vmware.api [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Waiting for the task: (returnval){ [ 1017.813810] env[62460]: value = "task-1314048" [ 1017.813810] env[62460]: _type = "Task" [ 1017.813810] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.822970] env[62460]: DEBUG oslo_vmware.api [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314048, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.003615] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1314047, 'name': CreateVM_Task, 'duration_secs': 0.40597} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.003795] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1018.004482] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.004667] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.004983] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1018.005259] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a70881c1-49c7-44aa-a6ba-9c9c37e35c9d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.009841] env[62460]: DEBUG oslo_vmware.api [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 1018.009841] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52eff2d4-924f-990f-10db-44350d91290b" [ 1018.009841] env[62460]: _type = "Task" [ 1018.009841] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.017350] env[62460]: DEBUG oslo_vmware.api [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52eff2d4-924f-990f-10db-44350d91290b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.278160] env[62460]: DEBUG oslo_vmware.api [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1314045, 'name': PowerOnVM_Task, 'duration_secs': 0.651221} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.278483] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1018.278774] env[62460]: INFO nova.compute.manager [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Took 7.20 seconds to spawn the instance on the hypervisor. [ 1018.279047] env[62460]: DEBUG nova.compute.manager [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1018.279892] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-379f346b-3f4f-4863-bee9-5ad112948bff {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.325258] env[62460]: DEBUG oslo_vmware.api [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314048, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.097279} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.325604] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1018.326426] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9313e421-39b6-4d04-9071-aca8c89b4406 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.349935] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac/c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1018.353578] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ba160b02-59e5-43ae-b860-187666dbece8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.375626] env[62460]: DEBUG oslo_vmware.api [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Waiting for the task: (returnval){ [ 1018.375626] env[62460]: value = "task-1314049" [ 1018.375626] env[62460]: _type = "Task" [ 1018.375626] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.385212] env[62460]: DEBUG oslo_vmware.api [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314049, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.519096] env[62460]: DEBUG oslo_vmware.api [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52eff2d4-924f-990f-10db-44350d91290b, 'name': SearchDatastore_Task, 'duration_secs': 0.009049} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.519415] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.519640] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1018.519876] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.520042] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.520235] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1018.520517] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d25802c8-54e2-4533-982f-2041f372c011 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.527873] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1018.528070] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1018.528789] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ee142558-de6c-4862-888d-7f5895efec1f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.533211] env[62460]: DEBUG oslo_vmware.api [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 1018.533211] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]5203ed9a-e72c-5a6e-fa3b-89de596510bc" [ 1018.533211] env[62460]: _type = "Task" [ 1018.533211] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.540079] env[62460]: DEBUG oslo_vmware.api [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5203ed9a-e72c-5a6e-fa3b-89de596510bc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.610613] env[62460]: DEBUG nova.network.neutron [req-67b9f0b7-de5e-4404-9234-00279b98ceeb req-8af8a5d1-c69a-4ed0-a1be-98b080d606df service nova] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Updated VIF entry in instance network info cache for port c591ec72-086f-44a0-8cf2-3e6719700505. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1018.611084] env[62460]: DEBUG nova.network.neutron [req-67b9f0b7-de5e-4404-9234-00279b98ceeb req-8af8a5d1-c69a-4ed0-a1be-98b080d606df service nova] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Updating instance_info_cache with network_info: [{"id": "c591ec72-086f-44a0-8cf2-3e6719700505", "address": "fa:16:3e:30:d9:88", "network": {"id": "b5cb4ee0-4562-45be-a25d-51c3a5932efd", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-823340443-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3d2d1c48ec14121a2e8c9b3f800a949", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55c757ac-f8b2-466d-b634-07dbd100b312", "external-id": "nsx-vlan-transportzone-159", "segmentation_id": 159, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc591ec72-08", "ovs_interfaceid": "c591ec72-086f-44a0-8cf2-3e6719700505", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1018.801548] env[62460]: INFO nova.compute.manager [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Took 14.28 seconds to build instance. [ 1018.886387] env[62460]: DEBUG oslo_vmware.api [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314049, 'name': ReconfigVM_Task, 'duration_secs': 0.275245} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.886672] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Reconfigured VM instance instance-00000066 to attach disk [datastore2] c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac/c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1018.887307] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e797bbcd-a399-4062-8072-869e4a9b04f9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.892686] env[62460]: DEBUG oslo_vmware.api [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Waiting for the task: (returnval){ [ 1018.892686] env[62460]: value = "task-1314050" [ 1018.892686] env[62460]: _type = "Task" [ 1018.892686] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.900644] env[62460]: DEBUG oslo_vmware.api [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314050, 'name': Rename_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.042654] env[62460]: DEBUG oslo_vmware.api [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5203ed9a-e72c-5a6e-fa3b-89de596510bc, 'name': SearchDatastore_Task, 'duration_secs': 0.007334} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.043600] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c9e555eb-c577-4cf1-aee6-c16558755188 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.048870] env[62460]: DEBUG oslo_vmware.api [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 1019.048870] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]520785bc-64d6-e84a-a8dc-70df89421612" [ 1019.048870] env[62460]: _type = "Task" [ 1019.048870] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.056354] env[62460]: DEBUG oslo_vmware.api [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]520785bc-64d6-e84a-a8dc-70df89421612, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.113660] env[62460]: DEBUG oslo_concurrency.lockutils [req-67b9f0b7-de5e-4404-9234-00279b98ceeb req-8af8a5d1-c69a-4ed0-a1be-98b080d606df service nova] Releasing lock "refresh_cache-00597ea8-88ee-448d-b5b5-2b29dc8683e9" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.304037] env[62460]: DEBUG oslo_concurrency.lockutils [None req-b98b22ca-3b06-4a0d-b1b5-0e1f25812f37 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "37e76fbf-fc84-4f63-ba7f-3515ccb376c1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.791s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.402364] env[62460]: DEBUG oslo_vmware.api [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314050, 'name': Rename_Task, 'duration_secs': 0.138876} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.403344] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1019.403344] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b06b4c37-ec90-43e3-873b-30ebd4b96449 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.410474] env[62460]: DEBUG oslo_vmware.api [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Waiting for the task: (returnval){ [ 1019.410474] env[62460]: value = "task-1314051" [ 1019.410474] env[62460]: _type = "Task" [ 1019.410474] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.420319] env[62460]: DEBUG oslo_vmware.api [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314051, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.559594] env[62460]: DEBUG oslo_vmware.api [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]520785bc-64d6-e84a-a8dc-70df89421612, 'name': SearchDatastore_Task, 'duration_secs': 0.009492} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.559883] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.560626] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] 00597ea8-88ee-448d-b5b5-2b29dc8683e9/00597ea8-88ee-448d-b5b5-2b29dc8683e9.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1019.560947] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7bfb9d93-e10b-4f34-a71c-c36ef4becb35 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.567992] env[62460]: DEBUG oslo_vmware.api [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 1019.567992] env[62460]: value = "task-1314052" [ 1019.567992] env[62460]: _type = "Task" [ 1019.567992] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.577163] env[62460]: DEBUG oslo_vmware.api [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314052, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.928598] env[62460]: DEBUG oslo_vmware.api [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314051, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.077474] env[62460]: DEBUG oslo_vmware.api [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314052, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.482497} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.077762] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore2] 00597ea8-88ee-448d-b5b5-2b29dc8683e9/00597ea8-88ee-448d-b5b5-2b29dc8683e9.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1020.077991] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1020.078450] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f9e9daee-7c80-44ae-8a5a-7a10ac1014b3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.086859] env[62460]: DEBUG oslo_vmware.api [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 1020.086859] env[62460]: value = "task-1314053" [ 1020.086859] env[62460]: _type = "Task" [ 1020.086859] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.096101] env[62460]: DEBUG oslo_vmware.api [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314053, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.422919] env[62460]: DEBUG oslo_vmware.api [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314051, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.596932] env[62460]: DEBUG oslo_vmware.api [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314053, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.191263} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.597735] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1020.599061] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-228ad466-4e5e-4d8e-9649-f66e4c542163 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.622600] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] 00597ea8-88ee-448d-b5b5-2b29dc8683e9/00597ea8-88ee-448d-b5b5-2b29dc8683e9.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1020.622970] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dad08dc0-d569-4f95-a023-699b76370456 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.643366] env[62460]: DEBUG oslo_vmware.api [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 1020.643366] env[62460]: value = "task-1314054" [ 1020.643366] env[62460]: _type = "Task" [ 1020.643366] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.652561] env[62460]: DEBUG oslo_vmware.api [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314054, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.799937] env[62460]: DEBUG oslo_concurrency.lockutils [None req-957f2070-23f9-4331-ab9c-55a694908063 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "37e76fbf-fc84-4f63-ba7f-3515ccb376c1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.800865] env[62460]: DEBUG oslo_concurrency.lockutils [None req-957f2070-23f9-4331-ab9c-55a694908063 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "37e76fbf-fc84-4f63-ba7f-3515ccb376c1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.800865] env[62460]: DEBUG oslo_concurrency.lockutils [None req-957f2070-23f9-4331-ab9c-55a694908063 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "37e76fbf-fc84-4f63-ba7f-3515ccb376c1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.800865] env[62460]: DEBUG oslo_concurrency.lockutils [None req-957f2070-23f9-4331-ab9c-55a694908063 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "37e76fbf-fc84-4f63-ba7f-3515ccb376c1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.800966] env[62460]: DEBUG oslo_concurrency.lockutils [None req-957f2070-23f9-4331-ab9c-55a694908063 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "37e76fbf-fc84-4f63-ba7f-3515ccb376c1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.803324] env[62460]: INFO nova.compute.manager [None req-957f2070-23f9-4331-ab9c-55a694908063 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Terminating instance [ 1020.807582] env[62460]: DEBUG nova.compute.manager [None req-957f2070-23f9-4331-ab9c-55a694908063 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1020.807800] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-957f2070-23f9-4331-ab9c-55a694908063 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1020.808671] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8918d3df-7a4b-4af0-8cda-9ed71208483c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.816430] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-957f2070-23f9-4331-ab9c-55a694908063 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1020.816685] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2ec00dc8-94cd-4d90-8f0a-ce7c4805fe52 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.825698] env[62460]: DEBUG oslo_vmware.api [None req-957f2070-23f9-4331-ab9c-55a694908063 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 1020.825698] env[62460]: value = "task-1314055" [ 1020.825698] env[62460]: _type = "Task" [ 1020.825698] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.836049] env[62460]: DEBUG oslo_vmware.api [None req-957f2070-23f9-4331-ab9c-55a694908063 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1314055, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.923129] env[62460]: DEBUG oslo_vmware.api [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314051, 'name': PowerOnVM_Task, 'duration_secs': 1.327197} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.923456] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1020.923662] env[62460]: INFO nova.compute.manager [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Took 7.67 seconds to spawn the instance on the hypervisor. [ 1020.923878] env[62460]: DEBUG nova.compute.manager [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1020.924672] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b21cb86-03c7-42a0-bbca-3ac8f5a37235 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.153994] env[62460]: DEBUG oslo_vmware.api [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314054, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.339295] env[62460]: DEBUG oslo_vmware.api [None req-957f2070-23f9-4331-ab9c-55a694908063 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1314055, 'name': PowerOffVM_Task, 'duration_secs': 0.199346} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.339764] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-957f2070-23f9-4331-ab9c-55a694908063 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1021.340071] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-957f2070-23f9-4331-ab9c-55a694908063 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1021.340427] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-19460871-cf82-4d35-b47e-89a6bfc55201 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.403020] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-957f2070-23f9-4331-ab9c-55a694908063 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1021.403020] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-957f2070-23f9-4331-ab9c-55a694908063 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Deleting contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1021.403020] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-957f2070-23f9-4331-ab9c-55a694908063 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Deleting the datastore file [datastore2] 37e76fbf-fc84-4f63-ba7f-3515ccb376c1 {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1021.403020] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bb978029-ffdd-48dc-a49b-441be0c57bd3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.410512] env[62460]: DEBUG oslo_vmware.api [None req-957f2070-23f9-4331-ab9c-55a694908063 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for the task: (returnval){ [ 1021.410512] env[62460]: value = "task-1314057" [ 1021.410512] env[62460]: _type = "Task" [ 1021.410512] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.423144] env[62460]: DEBUG oslo_vmware.api [None req-957f2070-23f9-4331-ab9c-55a694908063 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1314057, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.444918] env[62460]: INFO nova.compute.manager [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Took 13.80 seconds to build instance. [ 1021.654074] env[62460]: DEBUG oslo_vmware.api [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314054, 'name': ReconfigVM_Task, 'duration_secs': 0.785325} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.654411] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Reconfigured VM instance instance-00000067 to attach disk [datastore2] 00597ea8-88ee-448d-b5b5-2b29dc8683e9/00597ea8-88ee-448d-b5b5-2b29dc8683e9.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1021.655080] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-36cfdcba-a180-49df-92ef-2b1660aca64c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.661683] env[62460]: DEBUG oslo_vmware.api [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 1021.661683] env[62460]: value = "task-1314058" [ 1021.661683] env[62460]: _type = "Task" [ 1021.661683] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.668933] env[62460]: DEBUG oslo_vmware.api [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314058, 'name': Rename_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.920806] env[62460]: DEBUG oslo_vmware.api [None req-957f2070-23f9-4331-ab9c-55a694908063 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Task: {'id': task-1314057, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145319} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.920877] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-957f2070-23f9-4331-ab9c-55a694908063 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1021.921095] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-957f2070-23f9-4331-ab9c-55a694908063 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Deleted contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1021.921287] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-957f2070-23f9-4331-ab9c-55a694908063 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1021.921465] env[62460]: INFO nova.compute.manager [None req-957f2070-23f9-4331-ab9c-55a694908063 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1021.921713] env[62460]: DEBUG oslo.service.loopingcall [None req-957f2070-23f9-4331-ab9c-55a694908063 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1021.921918] env[62460]: DEBUG nova.compute.manager [-] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1021.922027] env[62460]: DEBUG nova.network.neutron [-] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1021.946368] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ac492d52-c271-417f-933e-9c2472f92b2d tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Lock "c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.306s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.112058] env[62460]: DEBUG oslo_concurrency.lockutils [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Acquiring lock "1f93e9e8-c589-4483-95f5-e0d62704562c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.112330] env[62460]: DEBUG oslo_concurrency.lockutils [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Lock "1f93e9e8-c589-4483-95f5-e0d62704562c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.173962] env[62460]: DEBUG oslo_vmware.api [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314058, 'name': Rename_Task, 'duration_secs': 0.236538} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.174346] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1022.174639] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a12ec276-a886-453d-b1ba-8172e4a15d7d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.182133] env[62460]: DEBUG oslo_vmware.api [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 1022.182133] env[62460]: value = "task-1314059" [ 1022.182133] env[62460]: _type = "Task" [ 1022.182133] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.189999] env[62460]: DEBUG oslo_vmware.api [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314059, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.266135] env[62460]: DEBUG nova.compute.manager [req-d43b0069-2cfa-4b17-9628-ba600277e094 req-c78db7a9-12cc-46a1-a4a3-b85f273a6ba0 service nova] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Received event network-vif-deleted-f516ed8d-249e-4717-bb01-d98f9e81cef5 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1022.266135] env[62460]: INFO nova.compute.manager [req-d43b0069-2cfa-4b17-9628-ba600277e094 req-c78db7a9-12cc-46a1-a4a3-b85f273a6ba0 service nova] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Neutron deleted interface f516ed8d-249e-4717-bb01-d98f9e81cef5; detaching it from the instance and deleting it from the info cache [ 1022.266407] env[62460]: DEBUG nova.network.neutron [req-d43b0069-2cfa-4b17-9628-ba600277e094 req-c78db7a9-12cc-46a1-a4a3-b85f273a6ba0 service nova] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1022.614948] env[62460]: DEBUG nova.compute.manager [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1022.692651] env[62460]: DEBUG oslo_vmware.api [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314059, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.737349] env[62460]: DEBUG nova.network.neutron [-] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1022.770063] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8d42b71e-a2c9-4724-b0af-23c9c15c4508 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.780648] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01b2f978-d043-4c7d-affb-6de0d26125be {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.795012] env[62460]: DEBUG oslo_concurrency.lockutils [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Acquiring lock "b618d776-de2d-497f-9eff-2e56043e44d4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.795217] env[62460]: DEBUG oslo_concurrency.lockutils [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Lock "b618d776-de2d-497f-9eff-2e56043e44d4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.808206] env[62460]: DEBUG nova.compute.manager [req-d43b0069-2cfa-4b17-9628-ba600277e094 req-c78db7a9-12cc-46a1-a4a3-b85f273a6ba0 service nova] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Detach interface failed, port_id=f516ed8d-249e-4717-bb01-d98f9e81cef5, reason: Instance 37e76fbf-fc84-4f63-ba7f-3515ccb376c1 could not be found. {{(pid=62460) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1023.134565] env[62460]: DEBUG oslo_concurrency.lockutils [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.134908] env[62460]: DEBUG oslo_concurrency.lockutils [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.136992] env[62460]: INFO nova.compute.claims [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1023.192454] env[62460]: DEBUG oslo_vmware.api [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314059, 'name': PowerOnVM_Task, 'duration_secs': 0.52497} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.192739] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1023.192948] env[62460]: INFO nova.compute.manager [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Took 7.74 seconds to spawn the instance on the hypervisor. [ 1023.193154] env[62460]: DEBUG nova.compute.manager [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1023.193956] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbfc92bb-9e44-41b2-aecf-db0f124df134 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.240268] env[62460]: INFO nova.compute.manager [-] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Took 1.32 seconds to deallocate network for instance. [ 1023.298033] env[62460]: DEBUG nova.compute.manager [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1023.710620] env[62460]: INFO nova.compute.manager [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Took 14.55 seconds to build instance. [ 1023.747697] env[62460]: DEBUG oslo_concurrency.lockutils [None req-957f2070-23f9-4331-ab9c-55a694908063 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.819403] env[62460]: DEBUG oslo_concurrency.lockutils [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.212369] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1c208641-76c8-4ba7-9cb1-8c62fa0854c4 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lock "00597ea8-88ee-448d-b5b5-2b29dc8683e9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.057s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.250366] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e843629-e61d-4777-bcff-f1eafb5305c3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.258873] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9be92b4f-b733-427f-9a0d-f8111c0e6c1c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.290747] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f00b6f49-6bee-4346-8bfe-a31c8ecd4255 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.298027] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebf79a59-9cc5-486a-883c-3a235d5e75b9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.311676] env[62460]: DEBUG nova.compute.provider_tree [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1024.625703] env[62460]: DEBUG nova.compute.manager [req-6fef4fa6-93c6-4bbc-87d4-c7a9a9548b8e req-114b9555-1146-4e84-9892-321f7eadfdf0 service nova] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Received event network-changed-c591ec72-086f-44a0-8cf2-3e6719700505 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1024.625912] env[62460]: DEBUG nova.compute.manager [req-6fef4fa6-93c6-4bbc-87d4-c7a9a9548b8e req-114b9555-1146-4e84-9892-321f7eadfdf0 service nova] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Refreshing instance network info cache due to event network-changed-c591ec72-086f-44a0-8cf2-3e6719700505. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1024.626192] env[62460]: DEBUG oslo_concurrency.lockutils [req-6fef4fa6-93c6-4bbc-87d4-c7a9a9548b8e req-114b9555-1146-4e84-9892-321f7eadfdf0 service nova] Acquiring lock "refresh_cache-00597ea8-88ee-448d-b5b5-2b29dc8683e9" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.626352] env[62460]: DEBUG oslo_concurrency.lockutils [req-6fef4fa6-93c6-4bbc-87d4-c7a9a9548b8e req-114b9555-1146-4e84-9892-321f7eadfdf0 service nova] Acquired lock "refresh_cache-00597ea8-88ee-448d-b5b5-2b29dc8683e9" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.626527] env[62460]: DEBUG nova.network.neutron [req-6fef4fa6-93c6-4bbc-87d4-c7a9a9548b8e req-114b9555-1146-4e84-9892-321f7eadfdf0 service nova] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Refreshing network info cache for port c591ec72-086f-44a0-8cf2-3e6719700505 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1024.814870] env[62460]: DEBUG nova.scheduler.client.report [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1025.319194] env[62460]: DEBUG oslo_concurrency.lockutils [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.184s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.320293] env[62460]: DEBUG oslo_concurrency.lockutils [None req-957f2070-23f9-4331-ab9c-55a694908063 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.573s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.320535] env[62460]: DEBUG nova.objects.instance [None req-957f2070-23f9-4331-ab9c-55a694908063 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lazy-loading 'resources' on Instance uuid 37e76fbf-fc84-4f63-ba7f-3515ccb376c1 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1025.353365] env[62460]: DEBUG nova.network.neutron [req-6fef4fa6-93c6-4bbc-87d4-c7a9a9548b8e req-114b9555-1146-4e84-9892-321f7eadfdf0 service nova] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Updated VIF entry in instance network info cache for port c591ec72-086f-44a0-8cf2-3e6719700505. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1025.353831] env[62460]: DEBUG nova.network.neutron [req-6fef4fa6-93c6-4bbc-87d4-c7a9a9548b8e req-114b9555-1146-4e84-9892-321f7eadfdf0 service nova] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Updating instance_info_cache with network_info: [{"id": "c591ec72-086f-44a0-8cf2-3e6719700505", "address": "fa:16:3e:30:d9:88", "network": {"id": "b5cb4ee0-4562-45be-a25d-51c3a5932efd", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-823340443-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.156", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3d2d1c48ec14121a2e8c9b3f800a949", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55c757ac-f8b2-466d-b634-07dbd100b312", "external-id": "nsx-vlan-transportzone-159", "segmentation_id": 159, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc591ec72-08", "ovs_interfaceid": "c591ec72-086f-44a0-8cf2-3e6719700505", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1025.823150] env[62460]: DEBUG oslo_concurrency.lockutils [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Acquiring lock "8356c6eb-e3d4-48de-b523-b1e7a964a908" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.823506] env[62460]: DEBUG oslo_concurrency.lockutils [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Lock "8356c6eb-e3d4-48de-b523-b1e7a964a908" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.856880] env[62460]: DEBUG oslo_concurrency.lockutils [req-6fef4fa6-93c6-4bbc-87d4-c7a9a9548b8e req-114b9555-1146-4e84-9892-321f7eadfdf0 service nova] Releasing lock "refresh_cache-00597ea8-88ee-448d-b5b5-2b29dc8683e9" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1025.918788] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b5168bc-d1af-41ca-8e9e-b5e7a02a92d9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.926240] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dec28ce2-834d-44c8-9d64-8d689c269c39 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.959232] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03746fda-a773-448a-9968-da61d3262d96 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.966477] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-112c0f9b-26a3-4c9f-b3c2-5eaee4a0d672 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.979109] env[62460]: DEBUG nova.compute.provider_tree [None req-957f2070-23f9-4331-ab9c-55a694908063 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1026.328307] env[62460]: DEBUG oslo_concurrency.lockutils [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Lock "8356c6eb-e3d4-48de-b523-b1e7a964a908" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.505s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.328959] env[62460]: DEBUG nova.compute.manager [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1026.482086] env[62460]: DEBUG nova.scheduler.client.report [None req-957f2070-23f9-4331-ab9c-55a694908063 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1026.834077] env[62460]: DEBUG nova.compute.utils [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1026.835793] env[62460]: DEBUG nova.compute.manager [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1026.836015] env[62460]: DEBUG nova.network.neutron [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1026.873644] env[62460]: DEBUG nova.policy [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '57d24ccd4efc4442899de66a587d37b1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9372526b4e4d4d5e9c0e99e5d1a93101', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 1026.986689] env[62460]: DEBUG oslo_concurrency.lockutils [None req-957f2070-23f9-4331-ab9c-55a694908063 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.666s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.989039] env[62460]: DEBUG oslo_concurrency.lockutils [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.170s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.990975] env[62460]: INFO nova.compute.claims [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1027.009542] env[62460]: INFO nova.scheduler.client.report [None req-957f2070-23f9-4331-ab9c-55a694908063 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Deleted allocations for instance 37e76fbf-fc84-4f63-ba7f-3515ccb376c1 [ 1027.137160] env[62460]: DEBUG nova.network.neutron [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Successfully created port: f8d517d2-c25a-4ee4-8a78-1318bc78ff43 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1027.339255] env[62460]: DEBUG nova.compute.manager [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1027.520424] env[62460]: DEBUG oslo_concurrency.lockutils [None req-957f2070-23f9-4331-ab9c-55a694908063 tempest-ServerDiskConfigTestJSON-82145115 tempest-ServerDiskConfigTestJSON-82145115-project-member] Lock "37e76fbf-fc84-4f63-ba7f-3515ccb376c1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.720s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.078799] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7c27060-066c-4b3f-8df0-f9beca872fcd {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.085998] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d22ae3e-7ecf-43d1-b6b8-d671eb1b5fea {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.115030] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59f5ec75-3cd4-47e5-bed8-14f03fe25d9e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.122300] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3db8e05a-157b-44e4-8517-1cad4b30575c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.135380] env[62460]: DEBUG nova.compute.provider_tree [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1028.349869] env[62460]: DEBUG nova.compute.manager [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1028.374487] env[62460]: DEBUG nova.virt.hardware [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1028.374771] env[62460]: DEBUG nova.virt.hardware [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1028.374973] env[62460]: DEBUG nova.virt.hardware [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1028.375214] env[62460]: DEBUG nova.virt.hardware [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1028.375377] env[62460]: DEBUG nova.virt.hardware [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1028.375536] env[62460]: DEBUG nova.virt.hardware [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1028.375750] env[62460]: DEBUG nova.virt.hardware [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1028.375919] env[62460]: DEBUG nova.virt.hardware [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1028.376104] env[62460]: DEBUG nova.virt.hardware [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1028.376280] env[62460]: DEBUG nova.virt.hardware [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1028.376460] env[62460]: DEBUG nova.virt.hardware [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1028.377339] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d043894-3408-4cd5-acaf-6925343139b9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.385745] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7faae1d-d4a6-4f46-b967-a17b16f23246 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.567198] env[62460]: DEBUG nova.compute.manager [req-cdfe5d38-17ea-4691-af86-2e869bcd90a2 req-b3cea8bb-0190-4638-b240-26ad9db4b21c service nova] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Received event network-vif-plugged-f8d517d2-c25a-4ee4-8a78-1318bc78ff43 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1028.567370] env[62460]: DEBUG oslo_concurrency.lockutils [req-cdfe5d38-17ea-4691-af86-2e869bcd90a2 req-b3cea8bb-0190-4638-b240-26ad9db4b21c service nova] Acquiring lock "1f93e9e8-c589-4483-95f5-e0d62704562c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.567586] env[62460]: DEBUG oslo_concurrency.lockutils [req-cdfe5d38-17ea-4691-af86-2e869bcd90a2 req-b3cea8bb-0190-4638-b240-26ad9db4b21c service nova] Lock "1f93e9e8-c589-4483-95f5-e0d62704562c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.567770] env[62460]: DEBUG oslo_concurrency.lockutils [req-cdfe5d38-17ea-4691-af86-2e869bcd90a2 req-b3cea8bb-0190-4638-b240-26ad9db4b21c service nova] Lock "1f93e9e8-c589-4483-95f5-e0d62704562c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.567946] env[62460]: DEBUG nova.compute.manager [req-cdfe5d38-17ea-4691-af86-2e869bcd90a2 req-b3cea8bb-0190-4638-b240-26ad9db4b21c service nova] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] No waiting events found dispatching network-vif-plugged-f8d517d2-c25a-4ee4-8a78-1318bc78ff43 {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1028.568411] env[62460]: WARNING nova.compute.manager [req-cdfe5d38-17ea-4691-af86-2e869bcd90a2 req-b3cea8bb-0190-4638-b240-26ad9db4b21c service nova] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Received unexpected event network-vif-plugged-f8d517d2-c25a-4ee4-8a78-1318bc78ff43 for instance with vm_state building and task_state spawning. [ 1028.638844] env[62460]: DEBUG nova.scheduler.client.report [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1028.741446] env[62460]: DEBUG nova.network.neutron [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Successfully updated port: f8d517d2-c25a-4ee4-8a78-1318bc78ff43 {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1029.144274] env[62460]: DEBUG oslo_concurrency.lockutils [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.155s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.144873] env[62460]: DEBUG nova.compute.manager [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1029.245553] env[62460]: DEBUG oslo_concurrency.lockutils [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Acquiring lock "refresh_cache-1f93e9e8-c589-4483-95f5-e0d62704562c" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1029.245553] env[62460]: DEBUG oslo_concurrency.lockutils [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Acquired lock "refresh_cache-1f93e9e8-c589-4483-95f5-e0d62704562c" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.245732] env[62460]: DEBUG nova.network.neutron [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1029.650367] env[62460]: DEBUG nova.compute.utils [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1029.655053] env[62460]: DEBUG nova.compute.manager [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1029.655338] env[62460]: DEBUG nova.network.neutron [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1029.696221] env[62460]: DEBUG nova.policy [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd2b0dac2c7fa4b60b547a3f081e27980', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4f13597d128c44e19b7f92f5ba1ba217', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 1029.791242] env[62460]: DEBUG nova.network.neutron [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1029.971556] env[62460]: DEBUG nova.network.neutron [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Successfully created port: fb790dca-5916-4987-a468-c6560299aa74 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1029.985377] env[62460]: DEBUG nova.network.neutron [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Updating instance_info_cache with network_info: [{"id": "f8d517d2-c25a-4ee4-8a78-1318bc78ff43", "address": "fa:16:3e:51:7d:1d", "network": {"id": "7fc525b1-a3a8-49ea-be9c-b2d48535ebfd", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-999090385-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9372526b4e4d4d5e9c0e99e5d1a93101", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa410d21-2141-45bb-8d0b-16c77304605f", "external-id": "nsx-vlan-transportzone-886", "segmentation_id": 886, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf8d517d2-c2", "ovs_interfaceid": "f8d517d2-c25a-4ee4-8a78-1318bc78ff43", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.155555] env[62460]: DEBUG nova.compute.manager [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1030.489046] env[62460]: DEBUG oslo_concurrency.lockutils [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Releasing lock "refresh_cache-1f93e9e8-c589-4483-95f5-e0d62704562c" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1030.489439] env[62460]: DEBUG nova.compute.manager [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Instance network_info: |[{"id": "f8d517d2-c25a-4ee4-8a78-1318bc78ff43", "address": "fa:16:3e:51:7d:1d", "network": {"id": "7fc525b1-a3a8-49ea-be9c-b2d48535ebfd", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-999090385-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9372526b4e4d4d5e9c0e99e5d1a93101", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa410d21-2141-45bb-8d0b-16c77304605f", "external-id": "nsx-vlan-transportzone-886", "segmentation_id": 886, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf8d517d2-c2", "ovs_interfaceid": "f8d517d2-c25a-4ee4-8a78-1318bc78ff43", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1030.489953] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:51:7d:1d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'aa410d21-2141-45bb-8d0b-16c77304605f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f8d517d2-c25a-4ee4-8a78-1318bc78ff43', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1030.498307] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Creating folder: Project (9372526b4e4d4d5e9c0e99e5d1a93101). Parent ref: group-v281134. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1030.498626] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3f4f04d4-d78b-41dc-a644-2f95149701bc {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.510360] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Created folder: Project (9372526b4e4d4d5e9c0e99e5d1a93101) in parent group-v281134. [ 1030.510513] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Creating folder: Instances. Parent ref: group-v281270. {{(pid=62460) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1030.510767] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-221f3eb4-419a-4c3a-9ed9-1a65979675ef {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.519635] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Created folder: Instances in parent group-v281270. [ 1030.519773] env[62460]: DEBUG oslo.service.loopingcall [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1030.520223] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1030.520223] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-90e4a3a2-0deb-487d-b741-3ee69b6fd9b7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.539216] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1030.539216] env[62460]: value = "task-1314062" [ 1030.539216] env[62460]: _type = "Task" [ 1030.539216] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.546939] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1314062, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.599634] env[62460]: DEBUG nova.compute.manager [req-bccceaab-fcab-41bb-bdff-8b722d0d661f req-ae32c409-b99a-434d-b1a3-ea2b13bde371 service nova] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Received event network-changed-f8d517d2-c25a-4ee4-8a78-1318bc78ff43 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1030.599985] env[62460]: DEBUG nova.compute.manager [req-bccceaab-fcab-41bb-bdff-8b722d0d661f req-ae32c409-b99a-434d-b1a3-ea2b13bde371 service nova] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Refreshing instance network info cache due to event network-changed-f8d517d2-c25a-4ee4-8a78-1318bc78ff43. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1030.600355] env[62460]: DEBUG oslo_concurrency.lockutils [req-bccceaab-fcab-41bb-bdff-8b722d0d661f req-ae32c409-b99a-434d-b1a3-ea2b13bde371 service nova] Acquiring lock "refresh_cache-1f93e9e8-c589-4483-95f5-e0d62704562c" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.600578] env[62460]: DEBUG oslo_concurrency.lockutils [req-bccceaab-fcab-41bb-bdff-8b722d0d661f req-ae32c409-b99a-434d-b1a3-ea2b13bde371 service nova] Acquired lock "refresh_cache-1f93e9e8-c589-4483-95f5-e0d62704562c" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.600777] env[62460]: DEBUG nova.network.neutron [req-bccceaab-fcab-41bb-bdff-8b722d0d661f req-ae32c409-b99a-434d-b1a3-ea2b13bde371 service nova] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Refreshing network info cache for port f8d517d2-c25a-4ee4-8a78-1318bc78ff43 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1031.050475] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1314062, 'name': CreateVM_Task, 'duration_secs': 0.282694} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.050690] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1031.051411] env[62460]: DEBUG oslo_concurrency.lockutils [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1031.051593] env[62460]: DEBUG oslo_concurrency.lockutils [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.051926] env[62460]: DEBUG oslo_concurrency.lockutils [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1031.052200] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dda6c2ea-d48a-436e-9875-ddfc2e5e6133 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.056770] env[62460]: DEBUG oslo_vmware.api [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Waiting for the task: (returnval){ [ 1031.056770] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52c9db8a-b9d3-8f93-6c47-3358bf63dd6c" [ 1031.056770] env[62460]: _type = "Task" [ 1031.056770] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.064262] env[62460]: DEBUG oslo_vmware.api [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52c9db8a-b9d3-8f93-6c47-3358bf63dd6c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.165061] env[62460]: DEBUG nova.compute.manager [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1031.190420] env[62460]: DEBUG nova.virt.hardware [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1031.190680] env[62460]: DEBUG nova.virt.hardware [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1031.190851] env[62460]: DEBUG nova.virt.hardware [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1031.191118] env[62460]: DEBUG nova.virt.hardware [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1031.192085] env[62460]: DEBUG nova.virt.hardware [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1031.192085] env[62460]: DEBUG nova.virt.hardware [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1031.192085] env[62460]: DEBUG nova.virt.hardware [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1031.192085] env[62460]: DEBUG nova.virt.hardware [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1031.192085] env[62460]: DEBUG nova.virt.hardware [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1031.192303] env[62460]: DEBUG nova.virt.hardware [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1031.192349] env[62460]: DEBUG nova.virt.hardware [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1031.193205] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be78bcaf-7fc2-430b-ba26-c2317e4cd466 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.202065] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b53bb3d-b52d-4c2b-8e5a-3094d73396d2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.365287] env[62460]: DEBUG nova.network.neutron [req-bccceaab-fcab-41bb-bdff-8b722d0d661f req-ae32c409-b99a-434d-b1a3-ea2b13bde371 service nova] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Updated VIF entry in instance network info cache for port f8d517d2-c25a-4ee4-8a78-1318bc78ff43. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1031.365648] env[62460]: DEBUG nova.network.neutron [req-bccceaab-fcab-41bb-bdff-8b722d0d661f req-ae32c409-b99a-434d-b1a3-ea2b13bde371 service nova] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Updating instance_info_cache with network_info: [{"id": "f8d517d2-c25a-4ee4-8a78-1318bc78ff43", "address": "fa:16:3e:51:7d:1d", "network": {"id": "7fc525b1-a3a8-49ea-be9c-b2d48535ebfd", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-999090385-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9372526b4e4d4d5e9c0e99e5d1a93101", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa410d21-2141-45bb-8d0b-16c77304605f", "external-id": "nsx-vlan-transportzone-886", "segmentation_id": 886, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf8d517d2-c2", "ovs_interfaceid": "f8d517d2-c25a-4ee4-8a78-1318bc78ff43", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1031.567371] env[62460]: DEBUG oslo_vmware.api [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52c9db8a-b9d3-8f93-6c47-3358bf63dd6c, 'name': SearchDatastore_Task, 'duration_secs': 0.009234} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.567772] env[62460]: DEBUG oslo_concurrency.lockutils [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1031.567948] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1031.568212] env[62460]: DEBUG oslo_concurrency.lockutils [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1031.568371] env[62460]: DEBUG oslo_concurrency.lockutils [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.568563] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1031.568843] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-90a262b5-b7ea-4e39-87e3-2257052b9b3c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.572637] env[62460]: DEBUG nova.network.neutron [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Successfully updated port: fb790dca-5916-4987-a468-c6560299aa74 {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1031.576977] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1031.577188] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1031.578130] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-216f8a98-fa55-44f7-bd20-bbf26a2367bb {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.583189] env[62460]: DEBUG oslo_vmware.api [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Waiting for the task: (returnval){ [ 1031.583189] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]5284b30d-68b9-d9dd-98f5-e095b67e52d1" [ 1031.583189] env[62460]: _type = "Task" [ 1031.583189] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.593725] env[62460]: DEBUG oslo_vmware.api [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5284b30d-68b9-d9dd-98f5-e095b67e52d1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.868138] env[62460]: DEBUG oslo_concurrency.lockutils [req-bccceaab-fcab-41bb-bdff-8b722d0d661f req-ae32c409-b99a-434d-b1a3-ea2b13bde371 service nova] Releasing lock "refresh_cache-1f93e9e8-c589-4483-95f5-e0d62704562c" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1032.079306] env[62460]: DEBUG oslo_concurrency.lockutils [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Acquiring lock "refresh_cache-b618d776-de2d-497f-9eff-2e56043e44d4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.079306] env[62460]: DEBUG oslo_concurrency.lockutils [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Acquired lock "refresh_cache-b618d776-de2d-497f-9eff-2e56043e44d4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.079582] env[62460]: DEBUG nova.network.neutron [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1032.094949] env[62460]: DEBUG oslo_vmware.api [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5284b30d-68b9-d9dd-98f5-e095b67e52d1, 'name': SearchDatastore_Task, 'duration_secs': 0.008658} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.095805] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d97d10a7-aa73-4d53-8f4a-1d469a84afa8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.101275] env[62460]: DEBUG oslo_vmware.api [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Waiting for the task: (returnval){ [ 1032.101275] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]526db3cf-0f6c-895f-e5d2-0a046e10a357" [ 1032.101275] env[62460]: _type = "Task" [ 1032.101275] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.109018] env[62460]: DEBUG oslo_vmware.api [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]526db3cf-0f6c-895f-e5d2-0a046e10a357, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.611733] env[62460]: DEBUG oslo_vmware.api [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]526db3cf-0f6c-895f-e5d2-0a046e10a357, 'name': SearchDatastore_Task, 'duration_secs': 0.009557} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.612053] env[62460]: DEBUG oslo_concurrency.lockutils [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1032.612343] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 1f93e9e8-c589-4483-95f5-e0d62704562c/1f93e9e8-c589-4483-95f5-e0d62704562c.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1032.612627] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a9e56303-4e28-4e68-9e8e-c7bd11333659 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.615745] env[62460]: DEBUG nova.network.neutron [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1032.619890] env[62460]: DEBUG oslo_vmware.api [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Waiting for the task: (returnval){ [ 1032.619890] env[62460]: value = "task-1314063" [ 1032.619890] env[62460]: _type = "Task" [ 1032.619890] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.628307] env[62460]: DEBUG oslo_vmware.api [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Task: {'id': task-1314063, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.637280] env[62460]: DEBUG nova.compute.manager [req-b7b784b7-724d-4780-8bcf-ddde29c11a05 req-5f29c626-a2a1-41c0-8c37-33bb977d06ff service nova] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Received event network-vif-plugged-fb790dca-5916-4987-a468-c6560299aa74 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1032.637466] env[62460]: DEBUG oslo_concurrency.lockutils [req-b7b784b7-724d-4780-8bcf-ddde29c11a05 req-5f29c626-a2a1-41c0-8c37-33bb977d06ff service nova] Acquiring lock "b618d776-de2d-497f-9eff-2e56043e44d4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.637685] env[62460]: DEBUG oslo_concurrency.lockutils [req-b7b784b7-724d-4780-8bcf-ddde29c11a05 req-5f29c626-a2a1-41c0-8c37-33bb977d06ff service nova] Lock "b618d776-de2d-497f-9eff-2e56043e44d4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.637860] env[62460]: DEBUG oslo_concurrency.lockutils [req-b7b784b7-724d-4780-8bcf-ddde29c11a05 req-5f29c626-a2a1-41c0-8c37-33bb977d06ff service nova] Lock "b618d776-de2d-497f-9eff-2e56043e44d4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.638049] env[62460]: DEBUG nova.compute.manager [req-b7b784b7-724d-4780-8bcf-ddde29c11a05 req-5f29c626-a2a1-41c0-8c37-33bb977d06ff service nova] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] No waiting events found dispatching network-vif-plugged-fb790dca-5916-4987-a468-c6560299aa74 {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1032.638355] env[62460]: WARNING nova.compute.manager [req-b7b784b7-724d-4780-8bcf-ddde29c11a05 req-5f29c626-a2a1-41c0-8c37-33bb977d06ff service nova] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Received unexpected event network-vif-plugged-fb790dca-5916-4987-a468-c6560299aa74 for instance with vm_state building and task_state spawning. [ 1032.638652] env[62460]: DEBUG nova.compute.manager [req-b7b784b7-724d-4780-8bcf-ddde29c11a05 req-5f29c626-a2a1-41c0-8c37-33bb977d06ff service nova] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Received event network-changed-fb790dca-5916-4987-a468-c6560299aa74 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1032.638899] env[62460]: DEBUG nova.compute.manager [req-b7b784b7-724d-4780-8bcf-ddde29c11a05 req-5f29c626-a2a1-41c0-8c37-33bb977d06ff service nova] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Refreshing instance network info cache due to event network-changed-fb790dca-5916-4987-a468-c6560299aa74. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1032.639109] env[62460]: DEBUG oslo_concurrency.lockutils [req-b7b784b7-724d-4780-8bcf-ddde29c11a05 req-5f29c626-a2a1-41c0-8c37-33bb977d06ff service nova] Acquiring lock "refresh_cache-b618d776-de2d-497f-9eff-2e56043e44d4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.773215] env[62460]: DEBUG nova.network.neutron [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Updating instance_info_cache with network_info: [{"id": "fb790dca-5916-4987-a468-c6560299aa74", "address": "fa:16:3e:db:b8:4b", "network": {"id": "3aa5e026-c4ee-4641-831b-a7e31b69adc7", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-702747461-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f13597d128c44e19b7f92f5ba1ba217", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b91b49a8-b849-4d0c-97f7-74fdcd88ae03", "external-id": "nsx-vlan-transportzone-406", "segmentation_id": 406, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfb790dca-59", "ovs_interfaceid": "fb790dca-5916-4987-a468-c6560299aa74", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1033.130484] env[62460]: DEBUG oslo_vmware.api [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Task: {'id': task-1314063, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.50864} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.130768] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 1f93e9e8-c589-4483-95f5-e0d62704562c/1f93e9e8-c589-4483-95f5-e0d62704562c.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1033.131096] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1033.131407] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-919a911e-7f79-47cc-91ee-97a215201858 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.137307] env[62460]: DEBUG oslo_vmware.api [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Waiting for the task: (returnval){ [ 1033.137307] env[62460]: value = "task-1314064" [ 1033.137307] env[62460]: _type = "Task" [ 1033.137307] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.146575] env[62460]: DEBUG oslo_vmware.api [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Task: {'id': task-1314064, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.276015] env[62460]: DEBUG oslo_concurrency.lockutils [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Releasing lock "refresh_cache-b618d776-de2d-497f-9eff-2e56043e44d4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.276402] env[62460]: DEBUG nova.compute.manager [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Instance network_info: |[{"id": "fb790dca-5916-4987-a468-c6560299aa74", "address": "fa:16:3e:db:b8:4b", "network": {"id": "3aa5e026-c4ee-4641-831b-a7e31b69adc7", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-702747461-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f13597d128c44e19b7f92f5ba1ba217", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b91b49a8-b849-4d0c-97f7-74fdcd88ae03", "external-id": "nsx-vlan-transportzone-406", "segmentation_id": 406, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfb790dca-59", "ovs_interfaceid": "fb790dca-5916-4987-a468-c6560299aa74", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1033.276738] env[62460]: DEBUG oslo_concurrency.lockutils [req-b7b784b7-724d-4780-8bcf-ddde29c11a05 req-5f29c626-a2a1-41c0-8c37-33bb977d06ff service nova] Acquired lock "refresh_cache-b618d776-de2d-497f-9eff-2e56043e44d4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.276981] env[62460]: DEBUG nova.network.neutron [req-b7b784b7-724d-4780-8bcf-ddde29c11a05 req-5f29c626-a2a1-41c0-8c37-33bb977d06ff service nova] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Refreshing network info cache for port fb790dca-5916-4987-a468-c6560299aa74 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1033.278248] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:db:b8:4b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b91b49a8-b849-4d0c-97f7-74fdcd88ae03', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fb790dca-5916-4987-a468-c6560299aa74', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1033.285910] env[62460]: DEBUG oslo.service.loopingcall [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1033.289106] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1033.289661] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-02158588-f81f-4174-a790-fb7afcf191ac {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.311091] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1033.311091] env[62460]: value = "task-1314065" [ 1033.311091] env[62460]: _type = "Task" [ 1033.311091] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.320649] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1314065, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.515344] env[62460]: DEBUG nova.network.neutron [req-b7b784b7-724d-4780-8bcf-ddde29c11a05 req-5f29c626-a2a1-41c0-8c37-33bb977d06ff service nova] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Updated VIF entry in instance network info cache for port fb790dca-5916-4987-a468-c6560299aa74. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1033.516056] env[62460]: DEBUG nova.network.neutron [req-b7b784b7-724d-4780-8bcf-ddde29c11a05 req-5f29c626-a2a1-41c0-8c37-33bb977d06ff service nova] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Updating instance_info_cache with network_info: [{"id": "fb790dca-5916-4987-a468-c6560299aa74", "address": "fa:16:3e:db:b8:4b", "network": {"id": "3aa5e026-c4ee-4641-831b-a7e31b69adc7", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-702747461-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f13597d128c44e19b7f92f5ba1ba217", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b91b49a8-b849-4d0c-97f7-74fdcd88ae03", "external-id": "nsx-vlan-transportzone-406", "segmentation_id": 406, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfb790dca-59", "ovs_interfaceid": "fb790dca-5916-4987-a468-c6560299aa74", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1033.648020] env[62460]: DEBUG oslo_vmware.api [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Task: {'id': task-1314064, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.197971} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.648385] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1033.649137] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac583c3a-ff85-4001-a575-1b4d0ce13345 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.672178] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Reconfiguring VM instance instance-00000068 to attach disk [datastore1] 1f93e9e8-c589-4483-95f5-e0d62704562c/1f93e9e8-c589-4483-95f5-e0d62704562c.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1033.672762] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7bd1580f-334d-46bd-85fc-1ed5f21fdf11 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.692480] env[62460]: DEBUG oslo_vmware.api [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Waiting for the task: (returnval){ [ 1033.692480] env[62460]: value = "task-1314066" [ 1033.692480] env[62460]: _type = "Task" [ 1033.692480] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.700548] env[62460]: DEBUG oslo_vmware.api [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Task: {'id': task-1314066, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.820242] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1314065, 'name': CreateVM_Task} progress is 25%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.018523] env[62460]: DEBUG oslo_concurrency.lockutils [req-b7b784b7-724d-4780-8bcf-ddde29c11a05 req-5f29c626-a2a1-41c0-8c37-33bb977d06ff service nova] Releasing lock "refresh_cache-b618d776-de2d-497f-9eff-2e56043e44d4" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1034.202206] env[62460]: DEBUG oslo_vmware.api [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Task: {'id': task-1314066, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.321922] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1314065, 'name': CreateVM_Task} progress is 25%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.707329] env[62460]: DEBUG oslo_vmware.api [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Task: {'id': task-1314066, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.824561] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1314065, 'name': CreateVM_Task, 'duration_secs': 1.363302} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.824703] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1034.825423] env[62460]: DEBUG oslo_concurrency.lockutils [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1034.825657] env[62460]: DEBUG oslo_concurrency.lockutils [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.826079] env[62460]: DEBUG oslo_concurrency.lockutils [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1034.826352] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8537751b-d4e0-42d6-b924-470a2e7ec4ee {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.831242] env[62460]: DEBUG oslo_vmware.api [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Waiting for the task: (returnval){ [ 1034.831242] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52581c6a-8700-3a87-abe6-88a2915ca553" [ 1034.831242] env[62460]: _type = "Task" [ 1034.831242] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.838379] env[62460]: DEBUG oslo_vmware.api [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52581c6a-8700-3a87-abe6-88a2915ca553, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.203506] env[62460]: DEBUG oslo_vmware.api [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Task: {'id': task-1314066, 'name': ReconfigVM_Task, 'duration_secs': 1.035962} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.203837] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Reconfigured VM instance instance-00000068 to attach disk [datastore1] 1f93e9e8-c589-4483-95f5-e0d62704562c/1f93e9e8-c589-4483-95f5-e0d62704562c.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1035.204561] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-59538d81-e813-40f6-9f46-5780e1c87b0d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.211249] env[62460]: DEBUG oslo_vmware.api [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Waiting for the task: (returnval){ [ 1035.211249] env[62460]: value = "task-1314067" [ 1035.211249] env[62460]: _type = "Task" [ 1035.211249] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.219117] env[62460]: DEBUG oslo_vmware.api [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Task: {'id': task-1314067, 'name': Rename_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.342068] env[62460]: DEBUG oslo_vmware.api [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52581c6a-8700-3a87-abe6-88a2915ca553, 'name': SearchDatastore_Task, 'duration_secs': 0.010739} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.342383] env[62460]: DEBUG oslo_concurrency.lockutils [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1035.342618] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1035.342883] env[62460]: DEBUG oslo_concurrency.lockutils [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1035.343059] env[62460]: DEBUG oslo_concurrency.lockutils [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.343256] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1035.343515] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-78345974-0183-4774-bf33-654b464554e3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.351353] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1035.351495] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1035.352171] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d9662981-6c4e-4414-8dac-25529470dac0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.356941] env[62460]: DEBUG oslo_vmware.api [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Waiting for the task: (returnval){ [ 1035.356941] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52f24aee-8698-4177-1e05-e60fe7486a4c" [ 1035.356941] env[62460]: _type = "Task" [ 1035.356941] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.363992] env[62460]: DEBUG oslo_vmware.api [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52f24aee-8698-4177-1e05-e60fe7486a4c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.722586] env[62460]: DEBUG oslo_vmware.api [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Task: {'id': task-1314067, 'name': Rename_Task, 'duration_secs': 0.131327} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.722996] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1035.723365] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-80000c27-f684-4ec1-9f79-f4f717c30ef1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.730150] env[62460]: DEBUG oslo_vmware.api [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Waiting for the task: (returnval){ [ 1035.730150] env[62460]: value = "task-1314068" [ 1035.730150] env[62460]: _type = "Task" [ 1035.730150] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.738276] env[62460]: DEBUG oslo_vmware.api [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Task: {'id': task-1314068, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.867071] env[62460]: DEBUG oslo_vmware.api [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52f24aee-8698-4177-1e05-e60fe7486a4c, 'name': SearchDatastore_Task, 'duration_secs': 0.007867} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.867879] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f32d3529-7e6d-4f68-97b8-c66cbc6189cc {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.872741] env[62460]: DEBUG oslo_vmware.api [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Waiting for the task: (returnval){ [ 1035.872741] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52468b67-884c-a196-8238-a5fa4eec6ca0" [ 1035.872741] env[62460]: _type = "Task" [ 1035.872741] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.880271] env[62460]: DEBUG oslo_vmware.api [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52468b67-884c-a196-8238-a5fa4eec6ca0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.240595] env[62460]: DEBUG oslo_vmware.api [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Task: {'id': task-1314068, 'name': PowerOnVM_Task, 'duration_secs': 0.431259} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.240825] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1036.241054] env[62460]: INFO nova.compute.manager [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Took 7.89 seconds to spawn the instance on the hypervisor. [ 1036.241252] env[62460]: DEBUG nova.compute.manager [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1036.242036] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43a41200-2dee-412f-8799-ddab7961fc14 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.383289] env[62460]: DEBUG oslo_vmware.api [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52468b67-884c-a196-8238-a5fa4eec6ca0, 'name': SearchDatastore_Task, 'duration_secs': 0.009617} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.383562] env[62460]: DEBUG oslo_concurrency.lockutils [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1036.383827] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] b618d776-de2d-497f-9eff-2e56043e44d4/b618d776-de2d-497f-9eff-2e56043e44d4.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1036.384112] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2a74fb7f-7569-4fc4-92d6-2d76eb6547ca {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.390649] env[62460]: DEBUG oslo_vmware.api [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Waiting for the task: (returnval){ [ 1036.390649] env[62460]: value = "task-1314069" [ 1036.390649] env[62460]: _type = "Task" [ 1036.390649] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.397943] env[62460]: DEBUG oslo_vmware.api [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314069, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.491236] env[62460]: DEBUG oslo_concurrency.lockutils [None req-69549bf0-40bb-4888-b57b-a3d7dc185ed7 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "61f4666e-0588-4a00-a969-2349148caef2" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.491466] env[62460]: DEBUG oslo_concurrency.lockutils [None req-69549bf0-40bb-4888-b57b-a3d7dc185ed7 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "61f4666e-0588-4a00-a969-2349148caef2" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.491658] env[62460]: DEBUG nova.compute.manager [None req-69549bf0-40bb-4888-b57b-a3d7dc185ed7 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1036.492531] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8476b406-e025-443f-a6cc-cf21f0f227c6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.498964] env[62460]: DEBUG nova.compute.manager [None req-69549bf0-40bb-4888-b57b-a3d7dc185ed7 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62460) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1036.499547] env[62460]: DEBUG nova.objects.instance [None req-69549bf0-40bb-4888-b57b-a3d7dc185ed7 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lazy-loading 'flavor' on Instance uuid 61f4666e-0588-4a00-a969-2349148caef2 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1036.763514] env[62460]: INFO nova.compute.manager [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Took 13.64 seconds to build instance. [ 1036.899824] env[62460]: DEBUG oslo_vmware.api [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314069, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.437901} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.900105] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] b618d776-de2d-497f-9eff-2e56043e44d4/b618d776-de2d-497f-9eff-2e56043e44d4.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1036.900364] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1036.900626] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5a56a673-67f3-4b96-b0e5-0aae86dd9be7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.907193] env[62460]: DEBUG oslo_vmware.api [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Waiting for the task: (returnval){ [ 1036.907193] env[62460]: value = "task-1314070" [ 1036.907193] env[62460]: _type = "Task" [ 1036.907193] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.915110] env[62460]: DEBUG oslo_vmware.api [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314070, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.004579] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-69549bf0-40bb-4888-b57b-a3d7dc185ed7 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1037.005209] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4b22583c-e8af-4ba5-875e-ec97104f8a86 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.011542] env[62460]: DEBUG oslo_vmware.api [None req-69549bf0-40bb-4888-b57b-a3d7dc185ed7 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1037.011542] env[62460]: value = "task-1314071" [ 1037.011542] env[62460]: _type = "Task" [ 1037.011542] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.020276] env[62460]: DEBUG oslo_vmware.api [None req-69549bf0-40bb-4888-b57b-a3d7dc185ed7 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314071, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.266438] env[62460]: DEBUG oslo_concurrency.lockutils [None req-346c279e-5fa0-4f01-8f87-7d96adde9570 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Lock "1f93e9e8-c589-4483-95f5-e0d62704562c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.154s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.417047] env[62460]: DEBUG oslo_vmware.api [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314070, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063237} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.417430] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1037.418293] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-404e9c7b-2186-4c2d-913f-aa4bc91db58c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.440462] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] b618d776-de2d-497f-9eff-2e56043e44d4/b618d776-de2d-497f-9eff-2e56043e44d4.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1037.440789] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1bc92101-fe5e-4c06-a03d-96691c340c84 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.460528] env[62460]: DEBUG oslo_vmware.api [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Waiting for the task: (returnval){ [ 1037.460528] env[62460]: value = "task-1314072" [ 1037.460528] env[62460]: _type = "Task" [ 1037.460528] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.468074] env[62460]: DEBUG oslo_vmware.api [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314072, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.521645] env[62460]: DEBUG oslo_vmware.api [None req-69549bf0-40bb-4888-b57b-a3d7dc185ed7 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314071, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.527241] env[62460]: DEBUG oslo_concurrency.lockutils [None req-16723df1-a113-4aba-aac8-73a0134c6be0 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Acquiring lock "1f93e9e8-c589-4483-95f5-e0d62704562c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.527507] env[62460]: DEBUG oslo_concurrency.lockutils [None req-16723df1-a113-4aba-aac8-73a0134c6be0 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Lock "1f93e9e8-c589-4483-95f5-e0d62704562c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.527738] env[62460]: DEBUG oslo_concurrency.lockutils [None req-16723df1-a113-4aba-aac8-73a0134c6be0 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Acquiring lock "1f93e9e8-c589-4483-95f5-e0d62704562c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.527949] env[62460]: DEBUG oslo_concurrency.lockutils [None req-16723df1-a113-4aba-aac8-73a0134c6be0 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Lock "1f93e9e8-c589-4483-95f5-e0d62704562c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.528148] env[62460]: DEBUG oslo_concurrency.lockutils [None req-16723df1-a113-4aba-aac8-73a0134c6be0 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Lock "1f93e9e8-c589-4483-95f5-e0d62704562c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.530407] env[62460]: INFO nova.compute.manager [None req-16723df1-a113-4aba-aac8-73a0134c6be0 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Terminating instance [ 1037.532369] env[62460]: DEBUG nova.compute.manager [None req-16723df1-a113-4aba-aac8-73a0134c6be0 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1037.532599] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-16723df1-a113-4aba-aac8-73a0134c6be0 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1037.533389] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31780c01-04b7-440a-a3cf-cd5b2928a34e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.540113] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-16723df1-a113-4aba-aac8-73a0134c6be0 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1037.540352] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7d140e1c-d0b7-4271-8263-17930888c8ef {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.547048] env[62460]: DEBUG oslo_vmware.api [None req-16723df1-a113-4aba-aac8-73a0134c6be0 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Waiting for the task: (returnval){ [ 1037.547048] env[62460]: value = "task-1314073" [ 1037.547048] env[62460]: _type = "Task" [ 1037.547048] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.554411] env[62460]: DEBUG oslo_vmware.api [None req-16723df1-a113-4aba-aac8-73a0134c6be0 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Task: {'id': task-1314073, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.971664] env[62460]: DEBUG oslo_vmware.api [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314072, 'name': ReconfigVM_Task, 'duration_secs': 0.381753} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.972042] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Reconfigured VM instance instance-00000069 to attach disk [datastore1] b618d776-de2d-497f-9eff-2e56043e44d4/b618d776-de2d-497f-9eff-2e56043e44d4.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1037.972722] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7566a2e2-73ea-4f98-b7d9-cd7b99239f76 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.978487] env[62460]: DEBUG oslo_vmware.api [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Waiting for the task: (returnval){ [ 1037.978487] env[62460]: value = "task-1314074" [ 1037.978487] env[62460]: _type = "Task" [ 1037.978487] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.987306] env[62460]: DEBUG oslo_vmware.api [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314074, 'name': Rename_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.020926] env[62460]: DEBUG oslo_vmware.api [None req-69549bf0-40bb-4888-b57b-a3d7dc185ed7 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314071, 'name': PowerOffVM_Task, 'duration_secs': 0.52096} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.021256] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-69549bf0-40bb-4888-b57b-a3d7dc185ed7 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1038.021488] env[62460]: DEBUG nova.compute.manager [None req-69549bf0-40bb-4888-b57b-a3d7dc185ed7 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1038.022529] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8135ffb-ede3-4712-9063-59a9b055090c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.057348] env[62460]: DEBUG oslo_vmware.api [None req-16723df1-a113-4aba-aac8-73a0134c6be0 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Task: {'id': task-1314073, 'name': PowerOffVM_Task, 'duration_secs': 0.162155} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.057644] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-16723df1-a113-4aba-aac8-73a0134c6be0 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1038.057870] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-16723df1-a113-4aba-aac8-73a0134c6be0 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1038.058184] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6449a75b-7e11-471c-b190-4946357a5a1e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.112226] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-16723df1-a113-4aba-aac8-73a0134c6be0 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1038.112458] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-16723df1-a113-4aba-aac8-73a0134c6be0 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Deleting contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1038.112655] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-16723df1-a113-4aba-aac8-73a0134c6be0 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Deleting the datastore file [datastore1] 1f93e9e8-c589-4483-95f5-e0d62704562c {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1038.112929] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c73eb562-dd3c-49d4-a33b-5d2856626057 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.118570] env[62460]: DEBUG oslo_vmware.api [None req-16723df1-a113-4aba-aac8-73a0134c6be0 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Waiting for the task: (returnval){ [ 1038.118570] env[62460]: value = "task-1314076" [ 1038.118570] env[62460]: _type = "Task" [ 1038.118570] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.125986] env[62460]: DEBUG oslo_vmware.api [None req-16723df1-a113-4aba-aac8-73a0134c6be0 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Task: {'id': task-1314076, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.488664] env[62460]: DEBUG oslo_vmware.api [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314074, 'name': Rename_Task, 'duration_secs': 0.123959} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.489008] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1038.489292] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-540f724c-a331-4731-9cc8-ff10fd4366ab {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.495561] env[62460]: DEBUG oslo_vmware.api [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Waiting for the task: (returnval){ [ 1038.495561] env[62460]: value = "task-1314077" [ 1038.495561] env[62460]: _type = "Task" [ 1038.495561] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.503223] env[62460]: DEBUG oslo_vmware.api [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314077, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.534866] env[62460]: DEBUG oslo_concurrency.lockutils [None req-69549bf0-40bb-4888-b57b-a3d7dc185ed7 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "61f4666e-0588-4a00-a969-2349148caef2" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.043s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.628902] env[62460]: DEBUG oslo_vmware.api [None req-16723df1-a113-4aba-aac8-73a0134c6be0 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Task: {'id': task-1314076, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.150977} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.629303] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-16723df1-a113-4aba-aac8-73a0134c6be0 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1038.629531] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-16723df1-a113-4aba-aac8-73a0134c6be0 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Deleted contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1038.629718] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-16723df1-a113-4aba-aac8-73a0134c6be0 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1038.629899] env[62460]: INFO nova.compute.manager [None req-16723df1-a113-4aba-aac8-73a0134c6be0 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1038.630176] env[62460]: DEBUG oslo.service.loopingcall [None req-16723df1-a113-4aba-aac8-73a0134c6be0 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1038.630375] env[62460]: DEBUG nova.compute.manager [-] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1038.630470] env[62460]: DEBUG nova.network.neutron [-] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1038.854244] env[62460]: DEBUG nova.objects.instance [None req-ef48ce38-f8ca-4350-bb8f-c4b3a10dd470 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lazy-loading 'flavor' on Instance uuid 61f4666e-0588-4a00-a969-2349148caef2 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1038.858817] env[62460]: DEBUG nova.compute.manager [req-82d36c3a-b20e-4949-8531-34f4b51d52ed req-3f585f38-85ac-4c21-9794-00698ac618a9 service nova] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Received event network-vif-deleted-f8d517d2-c25a-4ee4-8a78-1318bc78ff43 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1038.859044] env[62460]: INFO nova.compute.manager [req-82d36c3a-b20e-4949-8531-34f4b51d52ed req-3f585f38-85ac-4c21-9794-00698ac618a9 service nova] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Neutron deleted interface f8d517d2-c25a-4ee4-8a78-1318bc78ff43; detaching it from the instance and deleting it from the info cache [ 1038.859236] env[62460]: DEBUG nova.network.neutron [req-82d36c3a-b20e-4949-8531-34f4b51d52ed req-3f585f38-85ac-4c21-9794-00698ac618a9 service nova] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1039.005186] env[62460]: DEBUG oslo_vmware.api [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314077, 'name': PowerOnVM_Task, 'duration_secs': 0.440767} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.005507] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1039.005646] env[62460]: INFO nova.compute.manager [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Took 7.84 seconds to spawn the instance on the hypervisor. [ 1039.005834] env[62460]: DEBUG nova.compute.manager [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1039.006603] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb8b80ae-c3bc-475c-a066-7bb5b4861b02 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.336786] env[62460]: DEBUG nova.network.neutron [-] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1039.359879] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ef48ce38-f8ca-4350-bb8f-c4b3a10dd470 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "refresh_cache-61f4666e-0588-4a00-a969-2349148caef2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1039.360077] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ef48ce38-f8ca-4350-bb8f-c4b3a10dd470 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquired lock "refresh_cache-61f4666e-0588-4a00-a969-2349148caef2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1039.360262] env[62460]: DEBUG nova.network.neutron [None req-ef48ce38-f8ca-4350-bb8f-c4b3a10dd470 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1039.360447] env[62460]: DEBUG nova.objects.instance [None req-ef48ce38-f8ca-4350-bb8f-c4b3a10dd470 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lazy-loading 'info_cache' on Instance uuid 61f4666e-0588-4a00-a969-2349148caef2 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1039.362168] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8a0972c6-0a95-4e93-808a-f4c9d9be2f91 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.372023] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d13f8601-daab-422a-9331-f7af2a7fcd72 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.397239] env[62460]: DEBUG nova.compute.manager [req-82d36c3a-b20e-4949-8531-34f4b51d52ed req-3f585f38-85ac-4c21-9794-00698ac618a9 service nova] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Detach interface failed, port_id=f8d517d2-c25a-4ee4-8a78-1318bc78ff43, reason: Instance 1f93e9e8-c589-4483-95f5-e0d62704562c could not be found. {{(pid=62460) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1039.521572] env[62460]: INFO nova.compute.manager [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Took 15.72 seconds to build instance. [ 1039.778334] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e02e20fe-8321-457a-8905-015c162a24ce tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Acquiring lock "ddf1daf4-2772-48df-82c9-2192d012dedd" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.778334] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e02e20fe-8321-457a-8905-015c162a24ce tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lock "ddf1daf4-2772-48df-82c9-2192d012dedd" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.841235] env[62460]: INFO nova.compute.manager [-] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Took 1.21 seconds to deallocate network for instance. [ 1039.863770] env[62460]: DEBUG nova.objects.base [None req-ef48ce38-f8ca-4350-bb8f-c4b3a10dd470 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Object Instance<61f4666e-0588-4a00-a969-2349148caef2> lazy-loaded attributes: flavor,info_cache {{(pid=62460) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1040.023824] env[62460]: DEBUG oslo_concurrency.lockutils [None req-56ee5bc2-4eae-40ad-a519-e23036280027 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Lock "b618d776-de2d-497f-9eff-2e56043e44d4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.228s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.281224] env[62460]: DEBUG nova.compute.utils [None req-e02e20fe-8321-457a-8905-015c162a24ce tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1040.323171] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2f10531d-cbc9-47e2-8e4c-5cce49676ce2 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Acquiring lock "b618d776-de2d-497f-9eff-2e56043e44d4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.323484] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2f10531d-cbc9-47e2-8e4c-5cce49676ce2 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Lock "b618d776-de2d-497f-9eff-2e56043e44d4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.323710] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2f10531d-cbc9-47e2-8e4c-5cce49676ce2 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Acquiring lock "b618d776-de2d-497f-9eff-2e56043e44d4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.323906] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2f10531d-cbc9-47e2-8e4c-5cce49676ce2 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Lock "b618d776-de2d-497f-9eff-2e56043e44d4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.324098] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2f10531d-cbc9-47e2-8e4c-5cce49676ce2 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Lock "b618d776-de2d-497f-9eff-2e56043e44d4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.326117] env[62460]: INFO nova.compute.manager [None req-2f10531d-cbc9-47e2-8e4c-5cce49676ce2 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Terminating instance [ 1040.327898] env[62460]: DEBUG nova.compute.manager [None req-2f10531d-cbc9-47e2-8e4c-5cce49676ce2 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1040.328115] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2f10531d-cbc9-47e2-8e4c-5cce49676ce2 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1040.329053] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-402fa3bd-3d4b-4394-80eb-20e684133509 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.337053] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f10531d-cbc9-47e2-8e4c-5cce49676ce2 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1040.337311] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b11d4134-1bfa-447b-a37a-4a6498728ad6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.344301] env[62460]: DEBUG oslo_vmware.api [None req-2f10531d-cbc9-47e2-8e4c-5cce49676ce2 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Waiting for the task: (returnval){ [ 1040.344301] env[62460]: value = "task-1314078" [ 1040.344301] env[62460]: _type = "Task" [ 1040.344301] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.348125] env[62460]: DEBUG oslo_concurrency.lockutils [None req-16723df1-a113-4aba-aac8-73a0134c6be0 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.348365] env[62460]: DEBUG oslo_concurrency.lockutils [None req-16723df1-a113-4aba-aac8-73a0134c6be0 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.348608] env[62460]: DEBUG nova.objects.instance [None req-16723df1-a113-4aba-aac8-73a0134c6be0 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Lazy-loading 'resources' on Instance uuid 1f93e9e8-c589-4483-95f5-e0d62704562c {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1040.354213] env[62460]: DEBUG oslo_vmware.api [None req-2f10531d-cbc9-47e2-8e4c-5cce49676ce2 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314078, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.583783] env[62460]: DEBUG nova.network.neutron [None req-ef48ce38-f8ca-4350-bb8f-c4b3a10dd470 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Updating instance_info_cache with network_info: [{"id": "345c87ec-2bb1-420f-bc8c-845a795baad5", "address": "fa:16:3e:43:36:b0", "network": {"id": "1dfc278d-e89e-4caf-9cfb-c302106cf974", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-868970047-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eee8bf23869e4680aec16e37f972175f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap345c87ec-2b", "ovs_interfaceid": "345c87ec-2bb1-420f-bc8c-845a795baad5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.784561] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e02e20fe-8321-457a-8905-015c162a24ce tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lock "ddf1daf4-2772-48df-82c9-2192d012dedd" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.856447] env[62460]: DEBUG oslo_vmware.api [None req-2f10531d-cbc9-47e2-8e4c-5cce49676ce2 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314078, 'name': PowerOffVM_Task, 'duration_secs': 0.159293} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.857864] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f10531d-cbc9-47e2-8e4c-5cce49676ce2 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1040.857864] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2f10531d-cbc9-47e2-8e4c-5cce49676ce2 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1040.857864] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-15edc701-74cb-43a0-b87e-44a9a9952a47 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.916096] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2f10531d-cbc9-47e2-8e4c-5cce49676ce2 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1040.916344] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2f10531d-cbc9-47e2-8e4c-5cce49676ce2 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Deleting contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1040.916538] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-2f10531d-cbc9-47e2-8e4c-5cce49676ce2 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Deleting the datastore file [datastore1] b618d776-de2d-497f-9eff-2e56043e44d4 {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1040.918956] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c12105fc-ec7d-4cdb-b007-48a2c36dffab {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.925345] env[62460]: DEBUG oslo_vmware.api [None req-2f10531d-cbc9-47e2-8e4c-5cce49676ce2 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Waiting for the task: (returnval){ [ 1040.925345] env[62460]: value = "task-1314080" [ 1040.925345] env[62460]: _type = "Task" [ 1040.925345] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.934428] env[62460]: DEBUG oslo_vmware.api [None req-2f10531d-cbc9-47e2-8e4c-5cce49676ce2 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314080, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.940556] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86278551-234c-4f0a-8d9b-0ec7f630e449 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.947741] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c58724d3-827c-424f-bed5-ae2e18594634 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.978501] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-831395e0-5793-4d3b-a30a-11c416bca77a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.985588] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2b928ed-d7c4-4392-9296-e3c70e6cccb6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.998900] env[62460]: DEBUG nova.compute.provider_tree [None req-16723df1-a113-4aba-aac8-73a0134c6be0 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1041.086721] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ef48ce38-f8ca-4350-bb8f-c4b3a10dd470 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Releasing lock "refresh_cache-61f4666e-0588-4a00-a969-2349148caef2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1041.436991] env[62460]: DEBUG oslo_vmware.api [None req-2f10531d-cbc9-47e2-8e4c-5cce49676ce2 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314080, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.122798} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.437273] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-2f10531d-cbc9-47e2-8e4c-5cce49676ce2 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1041.437463] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2f10531d-cbc9-47e2-8e4c-5cce49676ce2 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Deleted contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1041.437644] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2f10531d-cbc9-47e2-8e4c-5cce49676ce2 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1041.437821] env[62460]: INFO nova.compute.manager [None req-2f10531d-cbc9-47e2-8e4c-5cce49676ce2 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1041.438083] env[62460]: DEBUG oslo.service.loopingcall [None req-2f10531d-cbc9-47e2-8e4c-5cce49676ce2 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1041.438287] env[62460]: DEBUG nova.compute.manager [-] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1041.438382] env[62460]: DEBUG nova.network.neutron [-] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1041.501654] env[62460]: DEBUG nova.scheduler.client.report [None req-16723df1-a113-4aba-aac8-73a0134c6be0 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1041.529373] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1041.529586] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1041.529986] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Starting heal instance info cache {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1041.529986] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Rebuilding the list of instances to heal {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1041.589909] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef48ce38-f8ca-4350-bb8f-c4b3a10dd470 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1041.590298] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b6058226-ef96-42a3-83c5-65f0329a93b6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.597767] env[62460]: DEBUG oslo_vmware.api [None req-ef48ce38-f8ca-4350-bb8f-c4b3a10dd470 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1041.597767] env[62460]: value = "task-1314081" [ 1041.597767] env[62460]: _type = "Task" [ 1041.597767] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.606079] env[62460]: DEBUG oslo_vmware.api [None req-ef48ce38-f8ca-4350-bb8f-c4b3a10dd470 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314081, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.681846] env[62460]: DEBUG nova.compute.manager [req-9dde040b-edf2-4d06-b021-b343edadadf1 req-6d763b69-f711-4f18-929a-d3391ac574d4 service nova] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Received event network-vif-deleted-fb790dca-5916-4987-a468-c6560299aa74 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1041.682165] env[62460]: INFO nova.compute.manager [req-9dde040b-edf2-4d06-b021-b343edadadf1 req-6d763b69-f711-4f18-929a-d3391ac574d4 service nova] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Neutron deleted interface fb790dca-5916-4987-a468-c6560299aa74; detaching it from the instance and deleting it from the info cache [ 1041.682368] env[62460]: DEBUG nova.network.neutron [req-9dde040b-edf2-4d06-b021-b343edadadf1 req-6d763b69-f711-4f18-929a-d3391ac574d4 service nova] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1041.842639] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e02e20fe-8321-457a-8905-015c162a24ce tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Acquiring lock "ddf1daf4-2772-48df-82c9-2192d012dedd" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.843447] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e02e20fe-8321-457a-8905-015c162a24ce tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lock "ddf1daf4-2772-48df-82c9-2192d012dedd" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.843447] env[62460]: INFO nova.compute.manager [None req-e02e20fe-8321-457a-8905-015c162a24ce tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Attaching volume 21eb1bc8-e4ea-4d71-b67d-460996c9ccf8 to /dev/sdb [ 1041.875946] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0d91af4-6933-4ad3-b508-e0967e24c3b7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.883219] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8db53ff0-6dfa-4517-b24d-934c85250e47 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.896095] env[62460]: DEBUG nova.virt.block_device [None req-e02e20fe-8321-457a-8905-015c162a24ce tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Updating existing volume attachment record: 7fe21e67-189f-4c20-8e83-1028ab4a8d9f {{(pid=62460) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1042.006250] env[62460]: DEBUG oslo_concurrency.lockutils [None req-16723df1-a113-4aba-aac8-73a0134c6be0 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.658s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.029078] env[62460]: INFO nova.scheduler.client.report [None req-16723df1-a113-4aba-aac8-73a0134c6be0 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Deleted allocations for instance 1f93e9e8-c589-4483-95f5-e0d62704562c [ 1042.033664] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Skipping network cache update for instance because it is being deleted. {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 1042.033890] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Acquiring lock "refresh_cache-61f4666e-0588-4a00-a969-2349148caef2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1042.034064] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Acquired lock "refresh_cache-61f4666e-0588-4a00-a969-2349148caef2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.034217] env[62460]: DEBUG nova.network.neutron [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Forcefully refreshing network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1042.034370] env[62460]: DEBUG nova.objects.instance [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lazy-loading 'info_cache' on Instance uuid 61f4666e-0588-4a00-a969-2349148caef2 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1042.107611] env[62460]: DEBUG oslo_vmware.api [None req-ef48ce38-f8ca-4350-bb8f-c4b3a10dd470 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314081, 'name': PowerOnVM_Task, 'duration_secs': 0.371178} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.107940] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef48ce38-f8ca-4350-bb8f-c4b3a10dd470 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1042.107940] env[62460]: DEBUG nova.compute.manager [None req-ef48ce38-f8ca-4350-bb8f-c4b3a10dd470 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1042.108826] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf87f8ad-9f6a-4980-82eb-47bad3e2bcd4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.159542] env[62460]: DEBUG nova.network.neutron [-] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1042.184850] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c6bff331-2f06-4888-bb47-737e22c32451 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.194602] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dd4df5b-845f-4230-b282-1c87ac2f04f1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.220969] env[62460]: DEBUG nova.compute.manager [req-9dde040b-edf2-4d06-b021-b343edadadf1 req-6d763b69-f711-4f18-929a-d3391ac574d4 service nova] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Detach interface failed, port_id=fb790dca-5916-4987-a468-c6560299aa74, reason: Instance b618d776-de2d-497f-9eff-2e56043e44d4 could not be found. {{(pid=62460) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1042.542271] env[62460]: DEBUG oslo_concurrency.lockutils [None req-16723df1-a113-4aba-aac8-73a0134c6be0 tempest-ServerGroupTestJSON-877786895 tempest-ServerGroupTestJSON-877786895-project-member] Lock "1f93e9e8-c589-4483-95f5-e0d62704562c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.015s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.662027] env[62460]: INFO nova.compute.manager [-] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Took 1.22 seconds to deallocate network for instance. [ 1043.167868] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2f10531d-cbc9-47e2-8e4c-5cce49676ce2 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.168215] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2f10531d-cbc9-47e2-8e4c-5cce49676ce2 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.168816] env[62460]: DEBUG nova.objects.instance [None req-2f10531d-cbc9-47e2-8e4c-5cce49676ce2 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Lazy-loading 'resources' on Instance uuid b618d776-de2d-497f-9eff-2e56043e44d4 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1043.766508] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1d6de0d-40e5-4b99-b759-71c31708fea0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.777247] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11fb0395-d26d-4fa6-b7d4-924556a06956 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.810957] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f463724a-66df-49b0-9231-0b290a4ba769 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.821964] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-491347de-548b-4447-877e-93b80ba427af {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.840255] env[62460]: DEBUG nova.compute.provider_tree [None req-2f10531d-cbc9-47e2-8e4c-5cce49676ce2 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1043.952141] env[62460]: DEBUG nova.network.neutron [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Updating instance_info_cache with network_info: [{"id": "345c87ec-2bb1-420f-bc8c-845a795baad5", "address": "fa:16:3e:43:36:b0", "network": {"id": "1dfc278d-e89e-4caf-9cfb-c302106cf974", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-868970047-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eee8bf23869e4680aec16e37f972175f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap345c87ec-2b", "ovs_interfaceid": "345c87ec-2bb1-420f-bc8c-845a795baad5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1044.343732] env[62460]: DEBUG nova.scheduler.client.report [None req-2f10531d-cbc9-47e2-8e4c-5cce49676ce2 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1044.454388] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Releasing lock "refresh_cache-61f4666e-0588-4a00-a969-2349148caef2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.454637] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Updated the network info_cache for instance {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1044.454860] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1044.455049] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1044.455216] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1044.455368] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1044.455513] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1044.455664] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1044.455797] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62460) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1044.455947] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1044.848820] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2f10531d-cbc9-47e2-8e4c-5cce49676ce2 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.680s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.869826] env[62460]: INFO nova.scheduler.client.report [None req-2f10531d-cbc9-47e2-8e4c-5cce49676ce2 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Deleted allocations for instance b618d776-de2d-497f-9eff-2e56043e44d4 [ 1044.959374] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.959600] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.959772] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.959923] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62460) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1044.960945] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98b86c81-2f87-4759-9e0f-ca03f1d56006 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.969195] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebf72ddd-8538-464e-aeb6-0acad4a2674e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.982909] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f1d8da7-af30-4bcd-a42f-00be22b3635c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.989440] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52e96917-69db-4adf-ac11-9469281d3bff {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.017304] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181023MB free_disk=135GB free_vcpus=48 pci_devices=None {{(pid=62460) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1045.017435] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.017621] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.377275] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2f10531d-cbc9-47e2-8e4c-5cce49676ce2 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Lock "b618d776-de2d-497f-9eff-2e56043e44d4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.054s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.044059] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 61f4666e-0588-4a00-a969-2349148caef2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1046.044235] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance ddf1daf4-2772-48df-82c9-2192d012dedd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1046.044363] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1046.044483] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 00597ea8-88ee-448d-b5b5-2b29dc8683e9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1046.044668] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=62460) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1046.044826] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=62460) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1046.100787] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed8c03fb-90e3-4165-9756-1ce0bccac103 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.108650] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea013e7c-248c-4070-a6c0-3d2cd3dc89bb {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.137778] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-577c1ad3-1830-4201-936d-3cb98105baa7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.144921] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2ff2950-cf00-427a-9b40-2f4f242a79f5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.157872] env[62460]: DEBUG nova.compute.provider_tree [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1046.442912] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-e02e20fe-8321-457a-8905-015c162a24ce tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Volume attach. Driver type: vmdk {{(pid=62460) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1046.443238] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-e02e20fe-8321-457a-8905-015c162a24ce tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281274', 'volume_id': '21eb1bc8-e4ea-4d71-b67d-460996c9ccf8', 'name': 'volume-21eb1bc8-e4ea-4d71-b67d-460996c9ccf8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ddf1daf4-2772-48df-82c9-2192d012dedd', 'attached_at': '', 'detached_at': '', 'volume_id': '21eb1bc8-e4ea-4d71-b67d-460996c9ccf8', 'serial': '21eb1bc8-e4ea-4d71-b67d-460996c9ccf8'} {{(pid=62460) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1046.444224] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb5ada63-effc-4cd6-a7a6-e6a35317ec54 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.462177] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da215162-89a3-4ecd-9508-4b1e378abf36 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.489786] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-e02e20fe-8321-457a-8905-015c162a24ce tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Reconfiguring VM instance instance-00000064 to attach disk [datastore2] volume-21eb1bc8-e4ea-4d71-b67d-460996c9ccf8/volume-21eb1bc8-e4ea-4d71-b67d-460996c9ccf8.vmdk or device None with type thin {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1046.490059] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-14fcfdd5-412a-4954-a557-0c0134b0ff47 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.510338] env[62460]: DEBUG oslo_vmware.api [None req-e02e20fe-8321-457a-8905-015c162a24ce tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Waiting for the task: (returnval){ [ 1046.510338] env[62460]: value = "task-1314084" [ 1046.510338] env[62460]: _type = "Task" [ 1046.510338] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.518216] env[62460]: DEBUG oslo_vmware.api [None req-e02e20fe-8321-457a-8905-015c162a24ce tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1314084, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.660831] env[62460]: DEBUG nova.scheduler.client.report [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1047.020291] env[62460]: DEBUG oslo_vmware.api [None req-e02e20fe-8321-457a-8905-015c162a24ce tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1314084, 'name': ReconfigVM_Task, 'duration_secs': 0.323839} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.020576] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-e02e20fe-8321-457a-8905-015c162a24ce tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Reconfigured VM instance instance-00000064 to attach disk [datastore2] volume-21eb1bc8-e4ea-4d71-b67d-460996c9ccf8/volume-21eb1bc8-e4ea-4d71-b67d-460996c9ccf8.vmdk or device None with type thin {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1047.025270] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-50b71f18-eff2-4249-9e89-8e2f35b922f7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.040086] env[62460]: DEBUG oslo_vmware.api [None req-e02e20fe-8321-457a-8905-015c162a24ce tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Waiting for the task: (returnval){ [ 1047.040086] env[62460]: value = "task-1314085" [ 1047.040086] env[62460]: _type = "Task" [ 1047.040086] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.048101] env[62460]: DEBUG oslo_vmware.api [None req-e02e20fe-8321-457a-8905-015c162a24ce tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1314085, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.166863] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62460) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1047.167075] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.149s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.549933] env[62460]: DEBUG oslo_vmware.api [None req-e02e20fe-8321-457a-8905-015c162a24ce tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1314085, 'name': ReconfigVM_Task, 'duration_secs': 0.128407} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.550312] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-e02e20fe-8321-457a-8905-015c162a24ce tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281274', 'volume_id': '21eb1bc8-e4ea-4d71-b67d-460996c9ccf8', 'name': 'volume-21eb1bc8-e4ea-4d71-b67d-460996c9ccf8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ddf1daf4-2772-48df-82c9-2192d012dedd', 'attached_at': '', 'detached_at': '', 'volume_id': '21eb1bc8-e4ea-4d71-b67d-460996c9ccf8', 'serial': '21eb1bc8-e4ea-4d71-b67d-460996c9ccf8'} {{(pid=62460) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1048.585198] env[62460]: DEBUG nova.objects.instance [None req-e02e20fe-8321-457a-8905-015c162a24ce tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lazy-loading 'flavor' on Instance uuid ddf1daf4-2772-48df-82c9-2192d012dedd {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1049.093071] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e02e20fe-8321-457a-8905-015c162a24ce tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lock "ddf1daf4-2772-48df-82c9-2192d012dedd" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.250s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.916335] env[62460]: DEBUG oslo_concurrency.lockutils [None req-848da483-fd64-4c91-9bd3-b1bf74296a13 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Acquiring lock "ddf1daf4-2772-48df-82c9-2192d012dedd" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.916585] env[62460]: DEBUG oslo_concurrency.lockutils [None req-848da483-fd64-4c91-9bd3-b1bf74296a13 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lock "ddf1daf4-2772-48df-82c9-2192d012dedd" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1050.420405] env[62460]: DEBUG nova.compute.utils [None req-848da483-fd64-4c91-9bd3-b1bf74296a13 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1050.923741] env[62460]: DEBUG oslo_concurrency.lockutils [None req-848da483-fd64-4c91-9bd3-b1bf74296a13 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lock "ddf1daf4-2772-48df-82c9-2192d012dedd" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.993987] env[62460]: DEBUG oslo_concurrency.lockutils [None req-848da483-fd64-4c91-9bd3-b1bf74296a13 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Acquiring lock "ddf1daf4-2772-48df-82c9-2192d012dedd" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.994320] env[62460]: DEBUG oslo_concurrency.lockutils [None req-848da483-fd64-4c91-9bd3-b1bf74296a13 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lock "ddf1daf4-2772-48df-82c9-2192d012dedd" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.994580] env[62460]: INFO nova.compute.manager [None req-848da483-fd64-4c91-9bd3-b1bf74296a13 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Attaching volume eab7708a-a9ca-44f1-b891-306cac04081e to /dev/sdc [ 1052.024589] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd545332-8258-4b78-a7d9-79921993e023 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.031923] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61edd388-6b71-4eec-b0e1-1760d086eea5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.044289] env[62460]: DEBUG nova.virt.block_device [None req-848da483-fd64-4c91-9bd3-b1bf74296a13 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Updating existing volume attachment record: 791ce49c-d795-4495-8815-abb56d026bb6 {{(pid=62460) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1055.414872] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Acquiring lock "c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.415275] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Lock "c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.415318] env[62460]: INFO nova.compute.manager [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Shelving [ 1055.922886] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1055.922886] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d0463010-0bc2-40ca-81d6-ea0946bdbafa {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.930052] env[62460]: DEBUG oslo_vmware.api [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Waiting for the task: (returnval){ [ 1055.930052] env[62460]: value = "task-1314088" [ 1055.930052] env[62460]: _type = "Task" [ 1055.930052] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.937832] env[62460]: DEBUG oslo_vmware.api [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314088, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.439658] env[62460]: DEBUG oslo_vmware.api [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314088, 'name': PowerOffVM_Task, 'duration_secs': 0.190823} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.440018] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1056.440764] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-126a2c9d-a9b6-4a62-9a97-3dd400d70100 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.458276] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce957b4b-831c-4cc9-b6d3-b7d614cac626 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.586466] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-848da483-fd64-4c91-9bd3-b1bf74296a13 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Volume attach. Driver type: vmdk {{(pid=62460) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1056.586779] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-848da483-fd64-4c91-9bd3-b1bf74296a13 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281275', 'volume_id': 'eab7708a-a9ca-44f1-b891-306cac04081e', 'name': 'volume-eab7708a-a9ca-44f1-b891-306cac04081e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ddf1daf4-2772-48df-82c9-2192d012dedd', 'attached_at': '', 'detached_at': '', 'volume_id': 'eab7708a-a9ca-44f1-b891-306cac04081e', 'serial': 'eab7708a-a9ca-44f1-b891-306cac04081e'} {{(pid=62460) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1056.587746] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38d34107-e578-4ee4-b283-a7a8723232fb {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.606670] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84d17275-6705-43c5-9e69-c73c6ea62c47 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.635668] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-848da483-fd64-4c91-9bd3-b1bf74296a13 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Reconfiguring VM instance instance-00000064 to attach disk [datastore2] volume-eab7708a-a9ca-44f1-b891-306cac04081e/volume-eab7708a-a9ca-44f1-b891-306cac04081e.vmdk or device None with type thin {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1056.635936] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9315bf3c-8064-4f0b-b282-b8b425c6452d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.655400] env[62460]: DEBUG oslo_vmware.api [None req-848da483-fd64-4c91-9bd3-b1bf74296a13 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Waiting for the task: (returnval){ [ 1056.655400] env[62460]: value = "task-1314089" [ 1056.655400] env[62460]: _type = "Task" [ 1056.655400] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.663438] env[62460]: DEBUG oslo_vmware.api [None req-848da483-fd64-4c91-9bd3-b1bf74296a13 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1314089, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.968297] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Creating Snapshot of the VM instance {{(pid=62460) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1056.969238] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-c3ea5905-f8d9-4274-a7cc-a21d1ec70dbe {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.977279] env[62460]: DEBUG oslo_vmware.api [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Waiting for the task: (returnval){ [ 1056.977279] env[62460]: value = "task-1314090" [ 1056.977279] env[62460]: _type = "Task" [ 1056.977279] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.987724] env[62460]: DEBUG oslo_vmware.api [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314090, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.166919] env[62460]: DEBUG oslo_vmware.api [None req-848da483-fd64-4c91-9bd3-b1bf74296a13 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1314089, 'name': ReconfigVM_Task, 'duration_secs': 0.349479} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.167231] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-848da483-fd64-4c91-9bd3-b1bf74296a13 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Reconfigured VM instance instance-00000064 to attach disk [datastore2] volume-eab7708a-a9ca-44f1-b891-306cac04081e/volume-eab7708a-a9ca-44f1-b891-306cac04081e.vmdk or device None with type thin {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1057.172365] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eb9b79c5-ae11-4648-847e-1bd8c1426d35 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.188555] env[62460]: DEBUG oslo_vmware.api [None req-848da483-fd64-4c91-9bd3-b1bf74296a13 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Waiting for the task: (returnval){ [ 1057.188555] env[62460]: value = "task-1314091" [ 1057.188555] env[62460]: _type = "Task" [ 1057.188555] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.197767] env[62460]: DEBUG oslo_vmware.api [None req-848da483-fd64-4c91-9bd3-b1bf74296a13 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1314091, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.487387] env[62460]: DEBUG oslo_vmware.api [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314090, 'name': CreateSnapshot_Task, 'duration_secs': 0.390108} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.487773] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Created Snapshot of the VM instance {{(pid=62460) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1057.488367] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86109330-f19e-41af-98d0-f066e4ddfe35 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.697849] env[62460]: DEBUG oslo_vmware.api [None req-848da483-fd64-4c91-9bd3-b1bf74296a13 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1314091, 'name': ReconfigVM_Task, 'duration_secs': 0.165236} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.698177] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-848da483-fd64-4c91-9bd3-b1bf74296a13 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281275', 'volume_id': 'eab7708a-a9ca-44f1-b891-306cac04081e', 'name': 'volume-eab7708a-a9ca-44f1-b891-306cac04081e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ddf1daf4-2772-48df-82c9-2192d012dedd', 'attached_at': '', 'detached_at': '', 'volume_id': 'eab7708a-a9ca-44f1-b891-306cac04081e', 'serial': 'eab7708a-a9ca-44f1-b891-306cac04081e'} {{(pid=62460) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1058.005607] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Creating linked-clone VM from snapshot {{(pid=62460) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1058.005948] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-11693622-be41-45e6-a6ce-5a303c156e72 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.014704] env[62460]: DEBUG oslo_vmware.api [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Waiting for the task: (returnval){ [ 1058.014704] env[62460]: value = "task-1314092" [ 1058.014704] env[62460]: _type = "Task" [ 1058.014704] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.022931] env[62460]: DEBUG oslo_vmware.api [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314092, 'name': CloneVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.525260] env[62460]: DEBUG oslo_vmware.api [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314092, 'name': CloneVM_Task} progress is 94%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.736036] env[62460]: DEBUG nova.objects.instance [None req-848da483-fd64-4c91-9bd3-b1bf74296a13 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lazy-loading 'flavor' on Instance uuid ddf1daf4-2772-48df-82c9-2192d012dedd {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1059.028848] env[62460]: DEBUG oslo_vmware.api [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314092, 'name': CloneVM_Task, 'duration_secs': 0.942424} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.029096] env[62460]: INFO nova.virt.vmwareapi.vmops [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Created linked-clone VM from snapshot [ 1059.029838] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eae87138-26b4-4ca9-b313-68d92baf3233 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.038080] env[62460]: DEBUG nova.virt.vmwareapi.images [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Uploading image 11c6008d-80c8-470b-9f2c-c42ba1856c06 {{(pid=62460) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1059.060012] env[62460]: DEBUG oslo_vmware.rw_handles [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1059.060012] env[62460]: value = "vm-281277" [ 1059.060012] env[62460]: _type = "VirtualMachine" [ 1059.060012] env[62460]: }. {{(pid=62460) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1059.060287] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-b578a481-045a-470b-9280-4c8c2a4b0b71 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.066914] env[62460]: DEBUG oslo_vmware.rw_handles [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Lease: (returnval){ [ 1059.066914] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52756b2a-b9d5-a75a-7f94-e2c03ab24a3b" [ 1059.066914] env[62460]: _type = "HttpNfcLease" [ 1059.066914] env[62460]: } obtained for exporting VM: (result){ [ 1059.066914] env[62460]: value = "vm-281277" [ 1059.066914] env[62460]: _type = "VirtualMachine" [ 1059.066914] env[62460]: }. {{(pid=62460) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1059.067230] env[62460]: DEBUG oslo_vmware.api [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Waiting for the lease: (returnval){ [ 1059.067230] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52756b2a-b9d5-a75a-7f94-e2c03ab24a3b" [ 1059.067230] env[62460]: _type = "HttpNfcLease" [ 1059.067230] env[62460]: } to be ready. {{(pid=62460) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1059.073437] env[62460]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1059.073437] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52756b2a-b9d5-a75a-7f94-e2c03ab24a3b" [ 1059.073437] env[62460]: _type = "HttpNfcLease" [ 1059.073437] env[62460]: } is initializing. {{(pid=62460) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1059.240904] env[62460]: DEBUG oslo_concurrency.lockutils [None req-848da483-fd64-4c91-9bd3-b1bf74296a13 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lock "ddf1daf4-2772-48df-82c9-2192d012dedd" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.246s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1059.524894] env[62460]: DEBUG oslo_concurrency.lockutils [None req-334b621f-8277-43ab-b999-1c5308258599 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Acquiring lock "ddf1daf4-2772-48df-82c9-2192d012dedd" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.525175] env[62460]: DEBUG oslo_concurrency.lockutils [None req-334b621f-8277-43ab-b999-1c5308258599 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lock "ddf1daf4-2772-48df-82c9-2192d012dedd" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.575363] env[62460]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1059.575363] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52756b2a-b9d5-a75a-7f94-e2c03ab24a3b" [ 1059.575363] env[62460]: _type = "HttpNfcLease" [ 1059.575363] env[62460]: } is ready. {{(pid=62460) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1059.575780] env[62460]: DEBUG oslo_vmware.rw_handles [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1059.575780] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52756b2a-b9d5-a75a-7f94-e2c03ab24a3b" [ 1059.575780] env[62460]: _type = "HttpNfcLease" [ 1059.575780] env[62460]: }. {{(pid=62460) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1059.576359] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68a71d5b-be09-47a9-a7e0-a045c04791d5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.583234] env[62460]: DEBUG oslo_vmware.rw_handles [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52558329-447d-01a6-fb16-3a21d89f7fcd/disk-0.vmdk from lease info. {{(pid=62460) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1059.583413] env[62460]: DEBUG oslo_vmware.rw_handles [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52558329-447d-01a6-fb16-3a21d89f7fcd/disk-0.vmdk for reading. {{(pid=62460) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1059.669624] env[62460]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-fa6da5f7-57e3-4f74-825f-3c9efd7ce996 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.028787] env[62460]: INFO nova.compute.manager [None req-334b621f-8277-43ab-b999-1c5308258599 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Detaching volume 21eb1bc8-e4ea-4d71-b67d-460996c9ccf8 [ 1060.064422] env[62460]: INFO nova.virt.block_device [None req-334b621f-8277-43ab-b999-1c5308258599 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Attempting to driver detach volume 21eb1bc8-e4ea-4d71-b67d-460996c9ccf8 from mountpoint /dev/sdb [ 1060.065153] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-334b621f-8277-43ab-b999-1c5308258599 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Volume detach. Driver type: vmdk {{(pid=62460) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1060.065153] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-334b621f-8277-43ab-b999-1c5308258599 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281274', 'volume_id': '21eb1bc8-e4ea-4d71-b67d-460996c9ccf8', 'name': 'volume-21eb1bc8-e4ea-4d71-b67d-460996c9ccf8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ddf1daf4-2772-48df-82c9-2192d012dedd', 'attached_at': '', 'detached_at': '', 'volume_id': '21eb1bc8-e4ea-4d71-b67d-460996c9ccf8', 'serial': '21eb1bc8-e4ea-4d71-b67d-460996c9ccf8'} {{(pid=62460) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1060.066037] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b51564b6-81da-45c0-8eb7-b300ddf1e3b8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.090515] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bc4fd41-76a2-45ef-bf95-c47bccdb9421 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.097813] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a825d933-b70b-4020-bc72-9d6bcbe62a66 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.122132] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd62907e-2b13-44e6-933a-7b5a9417f6ef {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.136856] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-334b621f-8277-43ab-b999-1c5308258599 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] The volume has not been displaced from its original location: [datastore2] volume-21eb1bc8-e4ea-4d71-b67d-460996c9ccf8/volume-21eb1bc8-e4ea-4d71-b67d-460996c9ccf8.vmdk. No consolidation needed. {{(pid=62460) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1060.142369] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-334b621f-8277-43ab-b999-1c5308258599 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Reconfiguring VM instance instance-00000064 to detach disk 2001 {{(pid=62460) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1060.142774] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a182abbd-903f-4eec-afa4-45b4b80bc7e7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.160826] env[62460]: DEBUG oslo_vmware.api [None req-334b621f-8277-43ab-b999-1c5308258599 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Waiting for the task: (returnval){ [ 1060.160826] env[62460]: value = "task-1314094" [ 1060.160826] env[62460]: _type = "Task" [ 1060.160826] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.169514] env[62460]: DEBUG oslo_vmware.api [None req-334b621f-8277-43ab-b999-1c5308258599 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1314094, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.670826] env[62460]: DEBUG oslo_vmware.api [None req-334b621f-8277-43ab-b999-1c5308258599 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1314094, 'name': ReconfigVM_Task, 'duration_secs': 0.22605} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.671243] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-334b621f-8277-43ab-b999-1c5308258599 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Reconfigured VM instance instance-00000064 to detach disk 2001 {{(pid=62460) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1060.676039] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1ca0d0c4-c67d-4d9a-8815-d3d809cd1cba {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.691198] env[62460]: DEBUG oslo_vmware.api [None req-334b621f-8277-43ab-b999-1c5308258599 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Waiting for the task: (returnval){ [ 1060.691198] env[62460]: value = "task-1314095" [ 1060.691198] env[62460]: _type = "Task" [ 1060.691198] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.699523] env[62460]: DEBUG oslo_vmware.api [None req-334b621f-8277-43ab-b999-1c5308258599 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1314095, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.201528] env[62460]: DEBUG oslo_vmware.api [None req-334b621f-8277-43ab-b999-1c5308258599 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1314095, 'name': ReconfigVM_Task, 'duration_secs': 0.139889} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.201956] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-334b621f-8277-43ab-b999-1c5308258599 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281274', 'volume_id': '21eb1bc8-e4ea-4d71-b67d-460996c9ccf8', 'name': 'volume-21eb1bc8-e4ea-4d71-b67d-460996c9ccf8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ddf1daf4-2772-48df-82c9-2192d012dedd', 'attached_at': '', 'detached_at': '', 'volume_id': '21eb1bc8-e4ea-4d71-b67d-460996c9ccf8', 'serial': '21eb1bc8-e4ea-4d71-b67d-460996c9ccf8'} {{(pid=62460) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1061.745088] env[62460]: DEBUG nova.objects.instance [None req-334b621f-8277-43ab-b999-1c5308258599 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lazy-loading 'flavor' on Instance uuid ddf1daf4-2772-48df-82c9-2192d012dedd {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1062.753016] env[62460]: DEBUG oslo_concurrency.lockutils [None req-334b621f-8277-43ab-b999-1c5308258599 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lock "ddf1daf4-2772-48df-82c9-2192d012dedd" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.228s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1062.769312] env[62460]: DEBUG oslo_concurrency.lockutils [None req-8d252c6d-ef0f-4c07-a995-3a8f5c017b75 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Acquiring lock "ddf1daf4-2772-48df-82c9-2192d012dedd" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1062.769695] env[62460]: DEBUG oslo_concurrency.lockutils [None req-8d252c6d-ef0f-4c07-a995-3a8f5c017b75 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lock "ddf1daf4-2772-48df-82c9-2192d012dedd" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.272426] env[62460]: INFO nova.compute.manager [None req-8d252c6d-ef0f-4c07-a995-3a8f5c017b75 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Detaching volume eab7708a-a9ca-44f1-b891-306cac04081e [ 1063.305072] env[62460]: INFO nova.virt.block_device [None req-8d252c6d-ef0f-4c07-a995-3a8f5c017b75 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Attempting to driver detach volume eab7708a-a9ca-44f1-b891-306cac04081e from mountpoint /dev/sdc [ 1063.305417] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d252c6d-ef0f-4c07-a995-3a8f5c017b75 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Volume detach. Driver type: vmdk {{(pid=62460) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1063.305750] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d252c6d-ef0f-4c07-a995-3a8f5c017b75 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281275', 'volume_id': 'eab7708a-a9ca-44f1-b891-306cac04081e', 'name': 'volume-eab7708a-a9ca-44f1-b891-306cac04081e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ddf1daf4-2772-48df-82c9-2192d012dedd', 'attached_at': '', 'detached_at': '', 'volume_id': 'eab7708a-a9ca-44f1-b891-306cac04081e', 'serial': 'eab7708a-a9ca-44f1-b891-306cac04081e'} {{(pid=62460) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1063.306761] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6de3b10f-f577-4d43-8f5f-4bc6e7a8f353 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.329835] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4928895-9d15-48e2-a769-2c0385091fb8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.337393] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ab097b7-c57c-444b-8783-3cf24ea68832 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.357463] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2055c48-5f97-4746-ace3-a28d57f48699 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.372128] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d252c6d-ef0f-4c07-a995-3a8f5c017b75 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] The volume has not been displaced from its original location: [datastore2] volume-eab7708a-a9ca-44f1-b891-306cac04081e/volume-eab7708a-a9ca-44f1-b891-306cac04081e.vmdk. No consolidation needed. {{(pid=62460) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1063.378636] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d252c6d-ef0f-4c07-a995-3a8f5c017b75 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Reconfiguring VM instance instance-00000064 to detach disk 2002 {{(pid=62460) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1063.379053] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aadd4b2c-bbad-41ec-88c1-7db86a4f4411 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.399050] env[62460]: DEBUG oslo_vmware.api [None req-8d252c6d-ef0f-4c07-a995-3a8f5c017b75 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Waiting for the task: (returnval){ [ 1063.399050] env[62460]: value = "task-1314096" [ 1063.399050] env[62460]: _type = "Task" [ 1063.399050] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.407953] env[62460]: DEBUG oslo_vmware.api [None req-8d252c6d-ef0f-4c07-a995-3a8f5c017b75 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1314096, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.909445] env[62460]: DEBUG oslo_vmware.api [None req-8d252c6d-ef0f-4c07-a995-3a8f5c017b75 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1314096, 'name': ReconfigVM_Task, 'duration_secs': 0.268793} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.909766] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d252c6d-ef0f-4c07-a995-3a8f5c017b75 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Reconfigured VM instance instance-00000064 to detach disk 2002 {{(pid=62460) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1063.914448] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-349a4cda-2493-482d-a01c-e1ea248aa612 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.929698] env[62460]: DEBUG oslo_vmware.api [None req-8d252c6d-ef0f-4c07-a995-3a8f5c017b75 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Waiting for the task: (returnval){ [ 1063.929698] env[62460]: value = "task-1314097" [ 1063.929698] env[62460]: _type = "Task" [ 1063.929698] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.937695] env[62460]: DEBUG oslo_vmware.api [None req-8d252c6d-ef0f-4c07-a995-3a8f5c017b75 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1314097, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.234609] env[62460]: DEBUG oslo_concurrency.lockutils [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquiring lock "00597ea8-88ee-448d-b5b5-2b29dc8683e9" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1064.234949] env[62460]: DEBUG oslo_concurrency.lockutils [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lock "00597ea8-88ee-448d-b5b5-2b29dc8683e9" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1064.235208] env[62460]: INFO nova.compute.manager [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Shelving [ 1064.440998] env[62460]: DEBUG oslo_vmware.api [None req-8d252c6d-ef0f-4c07-a995-3a8f5c017b75 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1314097, 'name': ReconfigVM_Task, 'duration_secs': 0.144301} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.441449] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d252c6d-ef0f-4c07-a995-3a8f5c017b75 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281275', 'volume_id': 'eab7708a-a9ca-44f1-b891-306cac04081e', 'name': 'volume-eab7708a-a9ca-44f1-b891-306cac04081e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ddf1daf4-2772-48df-82c9-2192d012dedd', 'attached_at': '', 'detached_at': '', 'volume_id': 'eab7708a-a9ca-44f1-b891-306cac04081e', 'serial': 'eab7708a-a9ca-44f1-b891-306cac04081e'} {{(pid=62460) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1064.744271] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1064.744492] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b3fa632b-6d17-45b2-ae5c-2e42968f657a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.751742] env[62460]: DEBUG oslo_vmware.api [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 1064.751742] env[62460]: value = "task-1314098" [ 1064.751742] env[62460]: _type = "Task" [ 1064.751742] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.761255] env[62460]: DEBUG oslo_vmware.api [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314098, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.989834] env[62460]: DEBUG nova.objects.instance [None req-8d252c6d-ef0f-4c07-a995-3a8f5c017b75 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lazy-loading 'flavor' on Instance uuid ddf1daf4-2772-48df-82c9-2192d012dedd {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1065.262839] env[62460]: DEBUG oslo_vmware.api [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314098, 'name': PowerOffVM_Task, 'duration_secs': 0.224926} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.263158] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1065.263944] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ad6d1ae-5c38-4b8e-93f7-2b878dbd43fe {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.282204] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bfad184-7a0b-4cc0-baf1-760d4f89a3b6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.793269] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Creating Snapshot of the VM instance {{(pid=62460) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1065.793715] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-0d49a347-ae22-44a9-860e-9a72f4a25866 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.801704] env[62460]: DEBUG oslo_vmware.api [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 1065.801704] env[62460]: value = "task-1314099" [ 1065.801704] env[62460]: _type = "Task" [ 1065.801704] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.809707] env[62460]: DEBUG oslo_vmware.api [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314099, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.998499] env[62460]: DEBUG oslo_concurrency.lockutils [None req-8d252c6d-ef0f-4c07-a995-3a8f5c017b75 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lock "ddf1daf4-2772-48df-82c9-2192d012dedd" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.228s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.312382] env[62460]: DEBUG oslo_vmware.api [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314099, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.813300] env[62460]: DEBUG oslo_vmware.api [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314099, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.166849] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c0c643b9-866c-4839-ba8b-fec692ddaab8 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Acquiring lock "ddf1daf4-2772-48df-82c9-2192d012dedd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.167189] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c0c643b9-866c-4839-ba8b-fec692ddaab8 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lock "ddf1daf4-2772-48df-82c9-2192d012dedd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.167189] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c0c643b9-866c-4839-ba8b-fec692ddaab8 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Acquiring lock "ddf1daf4-2772-48df-82c9-2192d012dedd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.167367] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c0c643b9-866c-4839-ba8b-fec692ddaab8 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lock "ddf1daf4-2772-48df-82c9-2192d012dedd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.168035] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c0c643b9-866c-4839-ba8b-fec692ddaab8 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lock "ddf1daf4-2772-48df-82c9-2192d012dedd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.169866] env[62460]: INFO nova.compute.manager [None req-c0c643b9-866c-4839-ba8b-fec692ddaab8 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Terminating instance [ 1067.172157] env[62460]: DEBUG nova.compute.manager [None req-c0c643b9-866c-4839-ba8b-fec692ddaab8 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1067.172387] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-c0c643b9-866c-4839-ba8b-fec692ddaab8 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1067.173328] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8827c1aa-e81c-499a-a864-f1813c8b8920 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.182132] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0c643b9-866c-4839-ba8b-fec692ddaab8 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1067.182405] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e52771d5-e190-4972-9a76-26495ac2ba18 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.188177] env[62460]: DEBUG oslo_vmware.api [None req-c0c643b9-866c-4839-ba8b-fec692ddaab8 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Waiting for the task: (returnval){ [ 1067.188177] env[62460]: value = "task-1314100" [ 1067.188177] env[62460]: _type = "Task" [ 1067.188177] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.196246] env[62460]: DEBUG oslo_vmware.api [None req-c0c643b9-866c-4839-ba8b-fec692ddaab8 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1314100, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.267574] env[62460]: DEBUG oslo_vmware.rw_handles [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52558329-447d-01a6-fb16-3a21d89f7fcd/disk-0.vmdk. {{(pid=62460) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1067.268589] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8aca22c-2ae7-4320-b2c7-e9d2578bf183 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.274504] env[62460]: DEBUG oslo_vmware.rw_handles [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52558329-447d-01a6-fb16-3a21d89f7fcd/disk-0.vmdk is in state: ready. {{(pid=62460) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1067.274680] env[62460]: ERROR oslo_vmware.rw_handles [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52558329-447d-01a6-fb16-3a21d89f7fcd/disk-0.vmdk due to incomplete transfer. [ 1067.274892] env[62460]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-7ec72971-97fa-4e08-a162-97e9a9890df3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.280848] env[62460]: DEBUG oslo_vmware.rw_handles [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52558329-447d-01a6-fb16-3a21d89f7fcd/disk-0.vmdk. {{(pid=62460) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1067.281052] env[62460]: DEBUG nova.virt.vmwareapi.images [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Uploaded image 11c6008d-80c8-470b-9f2c-c42ba1856c06 to the Glance image server {{(pid=62460) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1067.283315] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Destroying the VM {{(pid=62460) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1067.283539] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-91832a5c-bf7c-4c4f-bffa-3fcdd4c5227b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.289033] env[62460]: DEBUG oslo_vmware.api [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Waiting for the task: (returnval){ [ 1067.289033] env[62460]: value = "task-1314101" [ 1067.289033] env[62460]: _type = "Task" [ 1067.289033] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.298066] env[62460]: DEBUG oslo_vmware.api [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314101, 'name': Destroy_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.311870] env[62460]: DEBUG oslo_vmware.api [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314099, 'name': CreateSnapshot_Task, 'duration_secs': 1.279866} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.312139] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Created Snapshot of the VM instance {{(pid=62460) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1067.312848] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f30eefdb-3870-4fba-a10d-4b43c60ae3a7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.697644] env[62460]: DEBUG oslo_vmware.api [None req-c0c643b9-866c-4839-ba8b-fec692ddaab8 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1314100, 'name': PowerOffVM_Task, 'duration_secs': 0.166094} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.697902] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0c643b9-866c-4839-ba8b-fec692ddaab8 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1067.698096] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-c0c643b9-866c-4839-ba8b-fec692ddaab8 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1067.698356] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7277e3d7-99ed-4aff-b46c-86bca481dc3c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.754839] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-c0c643b9-866c-4839-ba8b-fec692ddaab8 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1067.755113] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-c0c643b9-866c-4839-ba8b-fec692ddaab8 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Deleting contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1067.755291] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0c643b9-866c-4839-ba8b-fec692ddaab8 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Deleting the datastore file [datastore1] ddf1daf4-2772-48df-82c9-2192d012dedd {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1067.755551] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dd36dd71-05db-4ab1-ba7a-80762333cebd {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.761521] env[62460]: DEBUG oslo_vmware.api [None req-c0c643b9-866c-4839-ba8b-fec692ddaab8 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Waiting for the task: (returnval){ [ 1067.761521] env[62460]: value = "task-1314103" [ 1067.761521] env[62460]: _type = "Task" [ 1067.761521] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.768545] env[62460]: DEBUG oslo_vmware.api [None req-c0c643b9-866c-4839-ba8b-fec692ddaab8 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1314103, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.797320] env[62460]: DEBUG oslo_vmware.api [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314101, 'name': Destroy_Task, 'duration_secs': 0.367204} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.797559] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Destroyed the VM [ 1067.797876] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Deleting Snapshot of the VM instance {{(pid=62460) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1067.798178] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-e32f3ff4-2f3a-4671-9162-25ed66ab1494 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.804276] env[62460]: DEBUG oslo_vmware.api [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Waiting for the task: (returnval){ [ 1067.804276] env[62460]: value = "task-1314104" [ 1067.804276] env[62460]: _type = "Task" [ 1067.804276] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.811414] env[62460]: DEBUG oslo_vmware.api [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314104, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.833016] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Creating linked-clone VM from snapshot {{(pid=62460) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1067.833330] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-13369f52-c11c-452c-985a-16b557c62e53 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.840563] env[62460]: DEBUG oslo_vmware.api [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 1067.840563] env[62460]: value = "task-1314105" [ 1067.840563] env[62460]: _type = "Task" [ 1067.840563] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.849631] env[62460]: DEBUG oslo_vmware.api [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314105, 'name': CloneVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.271031] env[62460]: DEBUG oslo_vmware.api [None req-c0c643b9-866c-4839-ba8b-fec692ddaab8 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Task: {'id': task-1314103, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.251875} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.271360] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0c643b9-866c-4839-ba8b-fec692ddaab8 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1068.271360] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-c0c643b9-866c-4839-ba8b-fec692ddaab8 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Deleted contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1068.271508] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-c0c643b9-866c-4839-ba8b-fec692ddaab8 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1068.271833] env[62460]: INFO nova.compute.manager [None req-c0c643b9-866c-4839-ba8b-fec692ddaab8 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1068.272139] env[62460]: DEBUG oslo.service.loopingcall [None req-c0c643b9-866c-4839-ba8b-fec692ddaab8 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1068.272379] env[62460]: DEBUG nova.compute.manager [-] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1068.272500] env[62460]: DEBUG nova.network.neutron [-] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1068.315786] env[62460]: DEBUG oslo_vmware.api [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314104, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.349848] env[62460]: DEBUG oslo_vmware.api [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314105, 'name': CloneVM_Task} progress is 94%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.725824] env[62460]: DEBUG nova.compute.manager [req-745bab31-bab1-450a-bcbc-9d1df5da5da8 req-aeb4efa6-4685-49b8-a79a-d4200bd2642c service nova] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Received event network-vif-deleted-5fe76c2c-4427-44ee-b05d-f0e6fbbd874e {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1068.726124] env[62460]: INFO nova.compute.manager [req-745bab31-bab1-450a-bcbc-9d1df5da5da8 req-aeb4efa6-4685-49b8-a79a-d4200bd2642c service nova] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Neutron deleted interface 5fe76c2c-4427-44ee-b05d-f0e6fbbd874e; detaching it from the instance and deleting it from the info cache [ 1068.726334] env[62460]: DEBUG nova.network.neutron [req-745bab31-bab1-450a-bcbc-9d1df5da5da8 req-aeb4efa6-4685-49b8-a79a-d4200bd2642c service nova] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1068.815505] env[62460]: DEBUG oslo_vmware.api [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314104, 'name': RemoveSnapshot_Task, 'duration_secs': 0.78228} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.816053] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Deleted Snapshot of the VM instance {{(pid=62460) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1068.816466] env[62460]: DEBUG nova.compute.manager [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1068.817286] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4d418d0-0efd-4d17-b9e5-0f4e72cd1c35 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.853632] env[62460]: DEBUG oslo_vmware.api [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314105, 'name': CloneVM_Task} progress is 94%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.197644] env[62460]: DEBUG nova.network.neutron [-] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1069.229296] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3602f35f-64bd-4ffb-be49-04c963855a7f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.238874] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3b78d0e-04d0-4ef2-a8e3-ce89818850a5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.266577] env[62460]: DEBUG nova.compute.manager [req-745bab31-bab1-450a-bcbc-9d1df5da5da8 req-aeb4efa6-4685-49b8-a79a-d4200bd2642c service nova] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Detach interface failed, port_id=5fe76c2c-4427-44ee-b05d-f0e6fbbd874e, reason: Instance ddf1daf4-2772-48df-82c9-2192d012dedd could not be found. {{(pid=62460) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1069.328711] env[62460]: INFO nova.compute.manager [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Shelve offloading [ 1069.331239] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1069.331526] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a618cef7-c509-43d8-9e8d-b6d5b745a03e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.338615] env[62460]: DEBUG oslo_vmware.api [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Waiting for the task: (returnval){ [ 1069.338615] env[62460]: value = "task-1314106" [ 1069.338615] env[62460]: _type = "Task" [ 1069.338615] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.352757] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] VM already powered off {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1069.352957] env[62460]: DEBUG nova.compute.manager [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1069.353247] env[62460]: DEBUG oslo_vmware.api [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314105, 'name': CloneVM_Task} progress is 100%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.353994] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48713b39-0a4c-4682-8c83-a187012d022a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.359549] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Acquiring lock "refresh_cache-c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1069.359748] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Acquired lock "refresh_cache-c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1069.359933] env[62460]: DEBUG nova.network.neutron [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1069.701496] env[62460]: INFO nova.compute.manager [-] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Took 1.43 seconds to deallocate network for instance. [ 1069.851845] env[62460]: DEBUG oslo_vmware.api [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314105, 'name': CloneVM_Task, 'duration_secs': 1.528789} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.852138] env[62460]: INFO nova.virt.vmwareapi.vmops [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Created linked-clone VM from snapshot [ 1069.852893] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e8017da-06ed-438f-91ba-479e999424db {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.859905] env[62460]: DEBUG nova.virt.vmwareapi.images [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Uploading image 5b45a9b7-790b-42da-bcad-cc40f09629fb {{(pid=62460) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1069.884299] env[62460]: DEBUG oslo_vmware.rw_handles [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1069.884299] env[62460]: value = "vm-281279" [ 1069.884299] env[62460]: _type = "VirtualMachine" [ 1069.884299] env[62460]: }. {{(pid=62460) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1069.884568] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-d00ce241-f1fb-4546-8f29-1878dbc9983f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.892726] env[62460]: DEBUG oslo_vmware.rw_handles [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lease: (returnval){ [ 1069.892726] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]527d010a-0b45-74eb-b3bf-d8d605475b3d" [ 1069.892726] env[62460]: _type = "HttpNfcLease" [ 1069.892726] env[62460]: } obtained for exporting VM: (result){ [ 1069.892726] env[62460]: value = "vm-281279" [ 1069.892726] env[62460]: _type = "VirtualMachine" [ 1069.892726] env[62460]: }. {{(pid=62460) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1069.892995] env[62460]: DEBUG oslo_vmware.api [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the lease: (returnval){ [ 1069.892995] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]527d010a-0b45-74eb-b3bf-d8d605475b3d" [ 1069.892995] env[62460]: _type = "HttpNfcLease" [ 1069.892995] env[62460]: } to be ready. {{(pid=62460) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1069.898863] env[62460]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1069.898863] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]527d010a-0b45-74eb-b3bf-d8d605475b3d" [ 1069.898863] env[62460]: _type = "HttpNfcLease" [ 1069.898863] env[62460]: } is initializing. {{(pid=62460) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1070.047476] env[62460]: DEBUG nova.network.neutron [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Updating instance_info_cache with network_info: [{"id": "d15647d8-f541-4169-bea9-0937db3301b1", "address": "fa:16:3e:28:b4:d2", "network": {"id": "3aa5e026-c4ee-4641-831b-a7e31b69adc7", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-702747461-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f13597d128c44e19b7f92f5ba1ba217", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b91b49a8-b849-4d0c-97f7-74fdcd88ae03", "external-id": "nsx-vlan-transportzone-406", "segmentation_id": 406, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd15647d8-f5", "ovs_interfaceid": "d15647d8-f541-4169-bea9-0937db3301b1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1070.208983] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c0c643b9-866c-4839-ba8b-fec692ddaab8 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.209328] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c0c643b9-866c-4839-ba8b-fec692ddaab8 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.209562] env[62460]: DEBUG nova.objects.instance [None req-c0c643b9-866c-4839-ba8b-fec692ddaab8 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lazy-loading 'resources' on Instance uuid ddf1daf4-2772-48df-82c9-2192d012dedd {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1070.400814] env[62460]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1070.400814] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]527d010a-0b45-74eb-b3bf-d8d605475b3d" [ 1070.400814] env[62460]: _type = "HttpNfcLease" [ 1070.400814] env[62460]: } is ready. {{(pid=62460) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1070.401206] env[62460]: DEBUG oslo_vmware.rw_handles [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1070.401206] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]527d010a-0b45-74eb-b3bf-d8d605475b3d" [ 1070.401206] env[62460]: _type = "HttpNfcLease" [ 1070.401206] env[62460]: }. {{(pid=62460) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1070.401876] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad756218-ca7d-4643-8920-eb5cab3f77be {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.410281] env[62460]: DEBUG oslo_vmware.rw_handles [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ef356b-b29a-c47e-22bf-72cb20bd7693/disk-0.vmdk from lease info. {{(pid=62460) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1070.410459] env[62460]: DEBUG oslo_vmware.rw_handles [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ef356b-b29a-c47e-22bf-72cb20bd7693/disk-0.vmdk for reading. {{(pid=62460) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1070.496055] env[62460]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-54a0c7cb-c9c9-400c-ab42-7b81083dbfad {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.550212] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Releasing lock "refresh_cache-c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1070.763659] env[62460]: DEBUG nova.compute.manager [req-3d5a1bdf-c7d2-4f07-bcf3-133745c8dae2 req-f2ae26c6-6b79-4133-bf09-3da347f5da70 service nova] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Received event network-vif-unplugged-d15647d8-f541-4169-bea9-0937db3301b1 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1070.763894] env[62460]: DEBUG oslo_concurrency.lockutils [req-3d5a1bdf-c7d2-4f07-bcf3-133745c8dae2 req-f2ae26c6-6b79-4133-bf09-3da347f5da70 service nova] Acquiring lock "c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.764168] env[62460]: DEBUG oslo_concurrency.lockutils [req-3d5a1bdf-c7d2-4f07-bcf3-133745c8dae2 req-f2ae26c6-6b79-4133-bf09-3da347f5da70 service nova] Lock "c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.764304] env[62460]: DEBUG oslo_concurrency.lockutils [req-3d5a1bdf-c7d2-4f07-bcf3-133745c8dae2 req-f2ae26c6-6b79-4133-bf09-3da347f5da70 service nova] Lock "c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.764536] env[62460]: DEBUG nova.compute.manager [req-3d5a1bdf-c7d2-4f07-bcf3-133745c8dae2 req-f2ae26c6-6b79-4133-bf09-3da347f5da70 service nova] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] No waiting events found dispatching network-vif-unplugged-d15647d8-f541-4169-bea9-0937db3301b1 {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1070.764671] env[62460]: WARNING nova.compute.manager [req-3d5a1bdf-c7d2-4f07-bcf3-133745c8dae2 req-f2ae26c6-6b79-4133-bf09-3da347f5da70 service nova] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Received unexpected event network-vif-unplugged-d15647d8-f541-4169-bea9-0937db3301b1 for instance with vm_state shelved and task_state shelving_offloading. [ 1070.783582] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d394e56a-1b3d-4913-adea-60ca11147c93 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.787950] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1070.788789] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd02e40b-77ed-4b0a-9f2a-ee552ab22b47 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.794469] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6710265d-7d7e-42e0-a5a0-d3a96f4587f2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.799955] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1070.800539] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a45497fb-bd52-498b-aed2-59884d481604 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.827140] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3fc9570-6055-4a11-a808-c0d538c58362 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.835918] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bde37970-2320-4991-9726-a29101af81f5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.849570] env[62460]: DEBUG nova.compute.provider_tree [None req-c0c643b9-866c-4839-ba8b-fec692ddaab8 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1070.855664] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1070.855917] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Deleting contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1070.856247] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Deleting the datastore file [datastore2] c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1070.856535] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d3fcd952-228c-4c4b-85ac-f7862166616b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.863035] env[62460]: DEBUG oslo_vmware.api [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Waiting for the task: (returnval){ [ 1070.863035] env[62460]: value = "task-1314109" [ 1070.863035] env[62460]: _type = "Task" [ 1070.863035] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.868601] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1070.868753] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Cleaning up deleted instances with incomplete migration {{(pid=62460) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 1070.872485] env[62460]: DEBUG oslo_vmware.api [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314109, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.353161] env[62460]: DEBUG nova.scheduler.client.report [None req-c0c643b9-866c-4839-ba8b-fec692ddaab8 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1071.375130] env[62460]: DEBUG oslo_vmware.api [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314109, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141382} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.376046] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1071.376046] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Deleted contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1071.376239] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1071.394202] env[62460]: INFO nova.scheduler.client.report [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Deleted allocations for instance c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac [ 1071.858661] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c0c643b9-866c-4839-ba8b-fec692ddaab8 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.649s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.876043] env[62460]: INFO nova.scheduler.client.report [None req-c0c643b9-866c-4839-ba8b-fec692ddaab8 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Deleted allocations for instance ddf1daf4-2772-48df-82c9-2192d012dedd [ 1071.899157] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1071.900013] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1071.900186] env[62460]: DEBUG nova.objects.instance [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Lazy-loading 'resources' on Instance uuid c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1072.387334] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c0c643b9-866c-4839-ba8b-fec692ddaab8 tempest-AttachVolumeTestJSON-638965176 tempest-AttachVolumeTestJSON-638965176-project-member] Lock "ddf1daf4-2772-48df-82c9-2192d012dedd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.220s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.403449] env[62460]: DEBUG nova.objects.instance [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Lazy-loading 'numa_topology' on Instance uuid c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1072.796799] env[62460]: DEBUG nova.compute.manager [req-68c50a17-b7c7-409f-acee-01816de3bdca req-f7d2fde4-a529-4ddc-87b3-2c3beb408648 service nova] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Received event network-changed-d15647d8-f541-4169-bea9-0937db3301b1 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1072.797174] env[62460]: DEBUG nova.compute.manager [req-68c50a17-b7c7-409f-acee-01816de3bdca req-f7d2fde4-a529-4ddc-87b3-2c3beb408648 service nova] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Refreshing instance network info cache due to event network-changed-d15647d8-f541-4169-bea9-0937db3301b1. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1072.797648] env[62460]: DEBUG oslo_concurrency.lockutils [req-68c50a17-b7c7-409f-acee-01816de3bdca req-f7d2fde4-a529-4ddc-87b3-2c3beb408648 service nova] Acquiring lock "refresh_cache-c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1072.797648] env[62460]: DEBUG oslo_concurrency.lockutils [req-68c50a17-b7c7-409f-acee-01816de3bdca req-f7d2fde4-a529-4ddc-87b3-2c3beb408648 service nova] Acquired lock "refresh_cache-c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1072.797938] env[62460]: DEBUG nova.network.neutron [req-68c50a17-b7c7-409f-acee-01816de3bdca req-f7d2fde4-a529-4ddc-87b3-2c3beb408648 service nova] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Refreshing network info cache for port d15647d8-f541-4169-bea9-0937db3301b1 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1072.906154] env[62460]: DEBUG nova.objects.base [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=62460) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1072.956057] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bac7251-97cb-4014-bff4-dba467e71244 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.963101] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c841604b-8ce5-4b44-b27d-8061a98f8916 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.992362] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9007824c-8423-44de-b0f1-bf24434e5e89 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.999331] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-117949ff-9525-419a-9207-9779e83fd6bc {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.014641] env[62460]: DEBUG nova.compute.provider_tree [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1073.300723] env[62460]: DEBUG oslo_concurrency.lockutils [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Acquiring lock "c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.377256] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1073.377417] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Starting heal instance info cache {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1073.509674] env[62460]: DEBUG nova.network.neutron [req-68c50a17-b7c7-409f-acee-01816de3bdca req-f7d2fde4-a529-4ddc-87b3-2c3beb408648 service nova] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Updated VIF entry in instance network info cache for port d15647d8-f541-4169-bea9-0937db3301b1. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1073.510088] env[62460]: DEBUG nova.network.neutron [req-68c50a17-b7c7-409f-acee-01816de3bdca req-f7d2fde4-a529-4ddc-87b3-2c3beb408648 service nova] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Updating instance_info_cache with network_info: [{"id": "d15647d8-f541-4169-bea9-0937db3301b1", "address": "fa:16:3e:28:b4:d2", "network": {"id": "3aa5e026-c4ee-4641-831b-a7e31b69adc7", "bridge": null, "label": "tempest-ServersNegativeTestJSON-702747461-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f13597d128c44e19b7f92f5ba1ba217", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapd15647d8-f5", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1073.518366] env[62460]: DEBUG nova.scheduler.client.report [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1074.014090] env[62460]: DEBUG oslo_concurrency.lockutils [req-68c50a17-b7c7-409f-acee-01816de3bdca req-f7d2fde4-a529-4ddc-87b3-2c3beb408648 service nova] Releasing lock "refresh_cache-c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1074.024256] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.125s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1074.385064] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Acquiring lock "refresh_cache-c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1074.385064] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Acquired lock "refresh_cache-c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1074.385064] env[62460]: DEBUG nova.network.neutron [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Forcefully refreshing network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1074.532082] env[62460]: DEBUG oslo_concurrency.lockutils [None req-ab981928-707c-4c7a-9a2c-4cf6e868dec4 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Lock "c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 19.117s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1074.532812] env[62460]: DEBUG oslo_concurrency.lockutils [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Lock "c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 1.232s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1074.532972] env[62460]: INFO nova.compute.manager [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Unshelving [ 1075.558475] env[62460]: DEBUG oslo_concurrency.lockutils [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1075.558828] env[62460]: DEBUG oslo_concurrency.lockutils [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1075.559468] env[62460]: DEBUG nova.objects.instance [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Lazy-loading 'pci_requests' on Instance uuid c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1075.613377] env[62460]: DEBUG nova.network.neutron [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Updating instance_info_cache with network_info: [{"id": "d15647d8-f541-4169-bea9-0937db3301b1", "address": "fa:16:3e:28:b4:d2", "network": {"id": "3aa5e026-c4ee-4641-831b-a7e31b69adc7", "bridge": null, "label": "tempest-ServersNegativeTestJSON-702747461-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f13597d128c44e19b7f92f5ba1ba217", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapd15647d8-f5", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1076.062881] env[62460]: DEBUG nova.objects.instance [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Lazy-loading 'numa_topology' on Instance uuid c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1076.116545] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Releasing lock "refresh_cache-c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1076.116734] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Updated the network info_cache for instance {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1076.116949] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1076.117133] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1076.117286] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1076.117440] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1076.117579] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1076.117726] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1076.117861] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62460) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1076.118028] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1076.118161] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Cleaning up deleted instances {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 1076.565463] env[62460]: INFO nova.compute.claims [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1076.631905] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] There are 45 instances to clean {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 1076.632196] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: b618d776-de2d-497f-9eff-2e56043e44d4] Instance has had 0 of 5 cleanup attempts {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1077.135475] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 1f93e9e8-c589-4483-95f5-e0d62704562c] Instance has had 0 of 5 cleanup attempts {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1077.314102] env[62460]: DEBUG oslo_vmware.rw_handles [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ef356b-b29a-c47e-22bf-72cb20bd7693/disk-0.vmdk. {{(pid=62460) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1077.315040] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb56ff3c-8369-4fc4-9545-8545671b4ba9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.321181] env[62460]: DEBUG oslo_vmware.rw_handles [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ef356b-b29a-c47e-22bf-72cb20bd7693/disk-0.vmdk is in state: ready. {{(pid=62460) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1077.321360] env[62460]: ERROR oslo_vmware.rw_handles [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ef356b-b29a-c47e-22bf-72cb20bd7693/disk-0.vmdk due to incomplete transfer. [ 1077.321609] env[62460]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-7666bb2d-90e2-4eb9-aee0-fb944ae3e29b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.327974] env[62460]: DEBUG oslo_vmware.rw_handles [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ef356b-b29a-c47e-22bf-72cb20bd7693/disk-0.vmdk. {{(pid=62460) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1077.328191] env[62460]: DEBUG nova.virt.vmwareapi.images [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Uploaded image 5b45a9b7-790b-42da-bcad-cc40f09629fb to the Glance image server {{(pid=62460) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1077.330484] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Destroying the VM {{(pid=62460) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1077.330715] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-4715ed2e-19d8-4553-a4cb-c726ca04acf7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.336303] env[62460]: DEBUG oslo_vmware.api [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 1077.336303] env[62460]: value = "task-1314112" [ 1077.336303] env[62460]: _type = "Task" [ 1077.336303] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.343341] env[62460]: DEBUG oslo_vmware.api [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314112, 'name': Destroy_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.621423] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42e48e3a-5161-4561-bc9c-009f4f2f7a1f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.628323] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f90f4ea0-259a-4e33-8857-812b6dfcb7e7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.656207] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 37e76fbf-fc84-4f63-ba7f-3515ccb376c1] Instance has had 0 of 5 cleanup attempts {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1077.658537] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2236cc1-974f-4ffc-a985-243affbfcb28 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.665512] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6fac1e9-b699-4b66-8a4b-0b13c8b47c80 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.680032] env[62460]: DEBUG nova.compute.provider_tree [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1077.846082] env[62460]: DEBUG oslo_vmware.api [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314112, 'name': Destroy_Task, 'duration_secs': 0.278681} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.846381] env[62460]: INFO nova.virt.vmwareapi.vm_util [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Destroyed the VM [ 1077.846639] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Deleting Snapshot of the VM instance {{(pid=62460) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1077.846896] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-b1bfd8d0-91d3-437b-9496-f57b08359cb2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.854489] env[62460]: DEBUG oslo_vmware.api [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 1077.854489] env[62460]: value = "task-1314113" [ 1077.854489] env[62460]: _type = "Task" [ 1077.854489] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.861880] env[62460]: DEBUG oslo_vmware.api [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314113, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.162580] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: ddf1daf4-2772-48df-82c9-2192d012dedd] Instance has had 0 of 5 cleanup attempts {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1078.183450] env[62460]: DEBUG nova.scheduler.client.report [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1078.364935] env[62460]: DEBUG oslo_vmware.api [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314113, 'name': RemoveSnapshot_Task, 'duration_secs': 0.403289} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.364935] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Deleted Snapshot of the VM instance {{(pid=62460) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1078.364935] env[62460]: DEBUG nova.compute.manager [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1078.365639] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcc74c91-e92e-4271-8171-a684f8b940de {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.416285] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3092cb46-2ad5-4a5c-bd84-f46ccc5fdfd4 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "61f4666e-0588-4a00-a969-2349148caef2" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1078.416516] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3092cb46-2ad5-4a5c-bd84-f46ccc5fdfd4 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "61f4666e-0588-4a00-a969-2349148caef2" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1078.416760] env[62460]: INFO nova.compute.manager [None req-3092cb46-2ad5-4a5c-bd84-f46ccc5fdfd4 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Rebooting instance [ 1078.667290] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 4c5e60fe-6366-4fb9-975f-52345de67478] Instance has had 0 of 5 cleanup attempts {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1078.689646] env[62460]: DEBUG oslo_concurrency.lockutils [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.131s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1078.735400] env[62460]: INFO nova.network.neutron [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Updating port d15647d8-f541-4169-bea9-0937db3301b1 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1078.877069] env[62460]: INFO nova.compute.manager [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Shelve offloading [ 1078.879276] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1078.879276] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6f323946-7eb0-4d9f-acc9-c0cad860448b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.886070] env[62460]: DEBUG oslo_vmware.api [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 1078.886070] env[62460]: value = "task-1314114" [ 1078.886070] env[62460]: _type = "Task" [ 1078.886070] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.894403] env[62460]: DEBUG oslo_vmware.api [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314114, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.939217] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3092cb46-2ad5-4a5c-bd84-f46ccc5fdfd4 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "refresh_cache-61f4666e-0588-4a00-a969-2349148caef2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1078.939413] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3092cb46-2ad5-4a5c-bd84-f46ccc5fdfd4 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquired lock "refresh_cache-61f4666e-0588-4a00-a969-2349148caef2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1078.939596] env[62460]: DEBUG nova.network.neutron [None req-3092cb46-2ad5-4a5c-bd84-f46ccc5fdfd4 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1079.171058] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: c772ecef-5cb5-4e69-a67c-0c487da4f1e3] Instance has had 0 of 5 cleanup attempts {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1079.395773] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] VM already powered off {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1079.395963] env[62460]: DEBUG nova.compute.manager [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1079.396677] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5211525-ec7a-4308-99b5-0ab7ce153210 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.403163] env[62460]: DEBUG oslo_concurrency.lockutils [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquiring lock "refresh_cache-00597ea8-88ee-448d-b5b5-2b29dc8683e9" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1079.403337] env[62460]: DEBUG oslo_concurrency.lockutils [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquired lock "refresh_cache-00597ea8-88ee-448d-b5b5-2b29dc8683e9" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1079.403525] env[62460]: DEBUG nova.network.neutron [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1079.675228] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 0d39a304-6a49-4d87-bfa2-4ba3b383578a] Instance has had 0 of 5 cleanup attempts {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1079.708823] env[62460]: DEBUG nova.network.neutron [None req-3092cb46-2ad5-4a5c-bd84-f46ccc5fdfd4 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Updating instance_info_cache with network_info: [{"id": "345c87ec-2bb1-420f-bc8c-845a795baad5", "address": "fa:16:3e:43:36:b0", "network": {"id": "1dfc278d-e89e-4caf-9cfb-c302106cf974", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-868970047-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eee8bf23869e4680aec16e37f972175f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap345c87ec-2b", "ovs_interfaceid": "345c87ec-2bb1-420f-bc8c-845a795baad5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1080.111586] env[62460]: DEBUG nova.network.neutron [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Updating instance_info_cache with network_info: [{"id": "c591ec72-086f-44a0-8cf2-3e6719700505", "address": "fa:16:3e:30:d9:88", "network": {"id": "b5cb4ee0-4562-45be-a25d-51c3a5932efd", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-823340443-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.156", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3d2d1c48ec14121a2e8c9b3f800a949", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55c757ac-f8b2-466d-b634-07dbd100b312", "external-id": "nsx-vlan-transportzone-159", "segmentation_id": 159, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc591ec72-08", "ovs_interfaceid": "c591ec72-086f-44a0-8cf2-3e6719700505", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1080.167192] env[62460]: DEBUG nova.compute.manager [req-ad8fbd0f-4d22-4743-ab5c-9502a4a2a51a req-544bc6bc-0176-4611-af85-c32006c01eee service nova] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Received event network-vif-plugged-d15647d8-f541-4169-bea9-0937db3301b1 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1080.167391] env[62460]: DEBUG oslo_concurrency.lockutils [req-ad8fbd0f-4d22-4743-ab5c-9502a4a2a51a req-544bc6bc-0176-4611-af85-c32006c01eee service nova] Acquiring lock "c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1080.167609] env[62460]: DEBUG oslo_concurrency.lockutils [req-ad8fbd0f-4d22-4743-ab5c-9502a4a2a51a req-544bc6bc-0176-4611-af85-c32006c01eee service nova] Lock "c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.167792] env[62460]: DEBUG oslo_concurrency.lockutils [req-ad8fbd0f-4d22-4743-ab5c-9502a4a2a51a req-544bc6bc-0176-4611-af85-c32006c01eee service nova] Lock "c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1080.167972] env[62460]: DEBUG nova.compute.manager [req-ad8fbd0f-4d22-4743-ab5c-9502a4a2a51a req-544bc6bc-0176-4611-af85-c32006c01eee service nova] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] No waiting events found dispatching network-vif-plugged-d15647d8-f541-4169-bea9-0937db3301b1 {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1080.168334] env[62460]: WARNING nova.compute.manager [req-ad8fbd0f-4d22-4743-ab5c-9502a4a2a51a req-544bc6bc-0176-4611-af85-c32006c01eee service nova] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Received unexpected event network-vif-plugged-d15647d8-f541-4169-bea9-0937db3301b1 for instance with vm_state shelved_offloaded and task_state spawning. [ 1080.178694] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: f0f66557-04ef-4a96-8a56-e640bfda3b36] Instance has had 0 of 5 cleanup attempts {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1080.210961] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3092cb46-2ad5-4a5c-bd84-f46ccc5fdfd4 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Releasing lock "refresh_cache-61f4666e-0588-4a00-a969-2349148caef2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1080.213791] env[62460]: DEBUG nova.compute.manager [None req-3092cb46-2ad5-4a5c-bd84-f46ccc5fdfd4 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1080.214675] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58f3bca6-1347-4283-8c72-45bf44ab6bf8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.298106] env[62460]: DEBUG oslo_concurrency.lockutils [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Acquiring lock "refresh_cache-c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1080.298334] env[62460]: DEBUG oslo_concurrency.lockutils [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Acquired lock "refresh_cache-c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1080.298585] env[62460]: DEBUG nova.network.neutron [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1080.616082] env[62460]: DEBUG oslo_concurrency.lockutils [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Releasing lock "refresh_cache-00597ea8-88ee-448d-b5b5-2b29dc8683e9" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1080.683061] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 290500b5-d5b4-43b6-a843-20dacf561f4b] Instance has had 0 of 5 cleanup attempts {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1080.937090] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1080.937975] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6389bbfa-877f-419e-9f83-f4fd20ad4570 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.945425] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1080.945664] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9e91da7b-a29a-42bf-b5c5-6d1aa19efbf9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.990143] env[62460]: DEBUG nova.network.neutron [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Updating instance_info_cache with network_info: [{"id": "d15647d8-f541-4169-bea9-0937db3301b1", "address": "fa:16:3e:28:b4:d2", "network": {"id": "3aa5e026-c4ee-4641-831b-a7e31b69adc7", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-702747461-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f13597d128c44e19b7f92f5ba1ba217", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b91b49a8-b849-4d0c-97f7-74fdcd88ae03", "external-id": "nsx-vlan-transportzone-406", "segmentation_id": 406, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd15647d8-f5", "ovs_interfaceid": "d15647d8-f541-4169-bea9-0937db3301b1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1081.010468] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1081.010708] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Deleting contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1081.010924] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Deleting the datastore file [datastore2] 00597ea8-88ee-448d-b5b5-2b29dc8683e9 {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1081.011390] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ca66c453-65c3-4c00-a744-1c34947a7030 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.018211] env[62460]: DEBUG oslo_vmware.api [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 1081.018211] env[62460]: value = "task-1314117" [ 1081.018211] env[62460]: _type = "Task" [ 1081.018211] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.026122] env[62460]: DEBUG oslo_vmware.api [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314117, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.186189] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 0a4cbbe0-5249-4f97-b77d-ebee5cbf0fea] Instance has had 0 of 5 cleanup attempts {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1081.229349] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb685732-9edc-4674-9fdd-a3aa4accc260 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.237136] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-3092cb46-2ad5-4a5c-bd84-f46ccc5fdfd4 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Doing hard reboot of VM {{(pid=62460) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 1081.237136] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-a57febc9-a37a-4ef6-9c7c-411ee61b108f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.242447] env[62460]: DEBUG oslo_vmware.api [None req-3092cb46-2ad5-4a5c-bd84-f46ccc5fdfd4 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1081.242447] env[62460]: value = "task-1314118" [ 1081.242447] env[62460]: _type = "Task" [ 1081.242447] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.250518] env[62460]: DEBUG oslo_vmware.api [None req-3092cb46-2ad5-4a5c-bd84-f46ccc5fdfd4 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314118, 'name': ResetVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.493373] env[62460]: DEBUG oslo_concurrency.lockutils [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Releasing lock "refresh_cache-c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1081.518740] env[62460]: DEBUG nova.virt.hardware [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='f1fc036e1206070121d47921b6211c02',container_format='bare',created_at=2024-10-18T09:08:57Z,direct_url=,disk_format='vmdk',id=11c6008d-80c8-470b-9f2c-c42ba1856c06,min_disk=1,min_ram=0,name='tempest-ServersNegativeTestJSON-server-2134996302-shelved',owner='4f13597d128c44e19b7f92f5ba1ba217',properties=ImageMetaProps,protected=,size=31661056,status='active',tags=,updated_at=2024-10-18T09:09:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1081.519928] env[62460]: DEBUG nova.virt.hardware [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1081.519928] env[62460]: DEBUG nova.virt.hardware [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1081.519928] env[62460]: DEBUG nova.virt.hardware [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1081.519928] env[62460]: DEBUG nova.virt.hardware [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1081.519928] env[62460]: DEBUG nova.virt.hardware [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1081.519928] env[62460]: DEBUG nova.virt.hardware [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1081.520184] env[62460]: DEBUG nova.virt.hardware [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1081.520214] env[62460]: DEBUG nova.virt.hardware [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1081.520359] env[62460]: DEBUG nova.virt.hardware [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1081.520563] env[62460]: DEBUG nova.virt.hardware [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1081.521484] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c43d8f5d-9a0f-40e1-8759-bcec2e80fcba {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.531358] env[62460]: DEBUG oslo_vmware.api [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314117, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.133134} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.533316] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1081.533511] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Deleted contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1081.533692] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1081.536656] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1aa7bd6b-04e1-43c9-a755-98c2f92fcf92 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.549756] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:28:b4:d2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b91b49a8-b849-4d0c-97f7-74fdcd88ae03', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd15647d8-f541-4169-bea9-0937db3301b1', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1081.557014] env[62460]: DEBUG oslo.service.loopingcall [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1081.557259] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1081.557544] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-51999252-5fc8-4aba-a471-59796dd548cd {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.575893] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1081.575893] env[62460]: value = "task-1314119" [ 1081.575893] env[62460]: _type = "Task" [ 1081.575893] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.582992] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1314119, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.656086] env[62460]: INFO nova.scheduler.client.report [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Deleted allocations for instance 00597ea8-88ee-448d-b5b5-2b29dc8683e9 [ 1081.689889] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: db331001-1744-4c42-a40c-87f3e60da4b8] Instance has had 0 of 5 cleanup attempts {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1081.752994] env[62460]: DEBUG oslo_vmware.api [None req-3092cb46-2ad5-4a5c-bd84-f46ccc5fdfd4 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314118, 'name': ResetVM_Task, 'duration_secs': 0.081608} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.753356] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-3092cb46-2ad5-4a5c-bd84-f46ccc5fdfd4 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Did hard reboot of VM {{(pid=62460) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 1081.753585] env[62460]: DEBUG nova.compute.manager [None req-3092cb46-2ad5-4a5c-bd84-f46ccc5fdfd4 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1081.754385] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e533669-4f4e-4cf3-b1c6-1f53d10ce839 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.085569] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1314119, 'name': CreateVM_Task} progress is 99%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.161047] env[62460]: DEBUG oslo_concurrency.lockutils [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1082.161363] env[62460]: DEBUG oslo_concurrency.lockutils [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.161603] env[62460]: DEBUG nova.objects.instance [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lazy-loading 'resources' on Instance uuid 00597ea8-88ee-448d-b5b5-2b29dc8683e9 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1082.192571] env[62460]: DEBUG nova.compute.manager [req-89538124-ff1f-44fa-8d36-5ab4f0f3bc15 req-0702b7c6-56ee-4df7-addc-563a2eb2efcf service nova] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Received event network-changed-d15647d8-f541-4169-bea9-0937db3301b1 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1082.192771] env[62460]: DEBUG nova.compute.manager [req-89538124-ff1f-44fa-8d36-5ab4f0f3bc15 req-0702b7c6-56ee-4df7-addc-563a2eb2efcf service nova] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Refreshing instance network info cache due to event network-changed-d15647d8-f541-4169-bea9-0937db3301b1. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1082.193030] env[62460]: DEBUG oslo_concurrency.lockutils [req-89538124-ff1f-44fa-8d36-5ab4f0f3bc15 req-0702b7c6-56ee-4df7-addc-563a2eb2efcf service nova] Acquiring lock "refresh_cache-c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1082.193188] env[62460]: DEBUG oslo_concurrency.lockutils [req-89538124-ff1f-44fa-8d36-5ab4f0f3bc15 req-0702b7c6-56ee-4df7-addc-563a2eb2efcf service nova] Acquired lock "refresh_cache-c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1082.193355] env[62460]: DEBUG nova.network.neutron [req-89538124-ff1f-44fa-8d36-5ab4f0f3bc15 req-0702b7c6-56ee-4df7-addc-563a2eb2efcf service nova] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Refreshing network info cache for port d15647d8-f541-4169-bea9-0937db3301b1 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1082.195453] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 003191d3-fe38-4bde-8d7c-75b4aad413b4] Instance has had 0 of 5 cleanup attempts {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1082.266799] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3092cb46-2ad5-4a5c-bd84-f46ccc5fdfd4 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "61f4666e-0588-4a00-a969-2349148caef2" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 3.850s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.587495] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1314119, 'name': CreateVM_Task} progress is 99%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.663977] env[62460]: DEBUG nova.objects.instance [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lazy-loading 'numa_topology' on Instance uuid 00597ea8-88ee-448d-b5b5-2b29dc8683e9 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1082.699939] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 9c63946c-f294-42ea-979e-68db26d88858] Instance has had 0 of 5 cleanup attempts {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1083.071649] env[62460]: DEBUG nova.network.neutron [req-89538124-ff1f-44fa-8d36-5ab4f0f3bc15 req-0702b7c6-56ee-4df7-addc-563a2eb2efcf service nova] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Updated VIF entry in instance network info cache for port d15647d8-f541-4169-bea9-0937db3301b1. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1083.072079] env[62460]: DEBUG nova.network.neutron [req-89538124-ff1f-44fa-8d36-5ab4f0f3bc15 req-0702b7c6-56ee-4df7-addc-563a2eb2efcf service nova] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Updating instance_info_cache with network_info: [{"id": "d15647d8-f541-4169-bea9-0937db3301b1", "address": "fa:16:3e:28:b4:d2", "network": {"id": "3aa5e026-c4ee-4641-831b-a7e31b69adc7", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-702747461-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f13597d128c44e19b7f92f5ba1ba217", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b91b49a8-b849-4d0c-97f7-74fdcd88ae03", "external-id": "nsx-vlan-transportzone-406", "segmentation_id": 406, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd15647d8-f5", "ovs_interfaceid": "d15647d8-f541-4169-bea9-0937db3301b1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1083.087286] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1314119, 'name': CreateVM_Task, 'duration_secs': 1.311235} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.087458] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1083.088121] env[62460]: DEBUG oslo_concurrency.lockutils [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c6008d-80c8-470b-9f2c-c42ba1856c06" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1083.088294] env[62460]: DEBUG oslo_concurrency.lockutils [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c6008d-80c8-470b-9f2c-c42ba1856c06" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.088721] env[62460]: DEBUG oslo_concurrency.lockutils [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11c6008d-80c8-470b-9f2c-c42ba1856c06" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1083.088989] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-42378494-9fdf-462b-a445-20a415a7c8ec {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.094065] env[62460]: DEBUG oslo_vmware.api [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Waiting for the task: (returnval){ [ 1083.094065] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52614e78-f1e2-1e1c-3e9c-62a49b0674b8" [ 1083.094065] env[62460]: _type = "Task" [ 1083.094065] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.103037] env[62460]: DEBUG oslo_vmware.api [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52614e78-f1e2-1e1c-3e9c-62a49b0674b8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.166227] env[62460]: DEBUG nova.objects.base [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Object Instance<00597ea8-88ee-448d-b5b5-2b29dc8683e9> lazy-loaded attributes: resources,numa_topology {{(pid=62460) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1083.203069] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 6a5934ed-30dc-4512-8cf2-1e4d9f20836c] Instance has had 0 of 5 cleanup attempts {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1083.218522] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d325fa51-47ef-4ee1-9400-a91de26138de {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.226875] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71e97b41-3f17-4e52-b1ea-2982d1d08ee8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.256774] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ea514f6-b679-4169-974b-a709c050edc8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.264584] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43d84476-567b-413d-b944-82a541408ca2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.278282] env[62460]: DEBUG nova.compute.provider_tree [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1083.574586] env[62460]: DEBUG oslo_concurrency.lockutils [req-89538124-ff1f-44fa-8d36-5ab4f0f3bc15 req-0702b7c6-56ee-4df7-addc-563a2eb2efcf service nova] Releasing lock "refresh_cache-c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1083.574894] env[62460]: DEBUG nova.compute.manager [req-89538124-ff1f-44fa-8d36-5ab4f0f3bc15 req-0702b7c6-56ee-4df7-addc-563a2eb2efcf service nova] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Received event network-vif-unplugged-c591ec72-086f-44a0-8cf2-3e6719700505 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1083.575142] env[62460]: DEBUG oslo_concurrency.lockutils [req-89538124-ff1f-44fa-8d36-5ab4f0f3bc15 req-0702b7c6-56ee-4df7-addc-563a2eb2efcf service nova] Acquiring lock "00597ea8-88ee-448d-b5b5-2b29dc8683e9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1083.575370] env[62460]: DEBUG oslo_concurrency.lockutils [req-89538124-ff1f-44fa-8d36-5ab4f0f3bc15 req-0702b7c6-56ee-4df7-addc-563a2eb2efcf service nova] Lock "00597ea8-88ee-448d-b5b5-2b29dc8683e9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1083.575540] env[62460]: DEBUG oslo_concurrency.lockutils [req-89538124-ff1f-44fa-8d36-5ab4f0f3bc15 req-0702b7c6-56ee-4df7-addc-563a2eb2efcf service nova] Lock "00597ea8-88ee-448d-b5b5-2b29dc8683e9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1083.575712] env[62460]: DEBUG nova.compute.manager [req-89538124-ff1f-44fa-8d36-5ab4f0f3bc15 req-0702b7c6-56ee-4df7-addc-563a2eb2efcf service nova] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] No waiting events found dispatching network-vif-unplugged-c591ec72-086f-44a0-8cf2-3e6719700505 {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1083.575889] env[62460]: WARNING nova.compute.manager [req-89538124-ff1f-44fa-8d36-5ab4f0f3bc15 req-0702b7c6-56ee-4df7-addc-563a2eb2efcf service nova] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Received unexpected event network-vif-unplugged-c591ec72-086f-44a0-8cf2-3e6719700505 for instance with vm_state shelved_offloaded and task_state None. [ 1083.576070] env[62460]: DEBUG nova.compute.manager [req-89538124-ff1f-44fa-8d36-5ab4f0f3bc15 req-0702b7c6-56ee-4df7-addc-563a2eb2efcf service nova] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Received event network-changed-c591ec72-086f-44a0-8cf2-3e6719700505 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1083.576236] env[62460]: DEBUG nova.compute.manager [req-89538124-ff1f-44fa-8d36-5ab4f0f3bc15 req-0702b7c6-56ee-4df7-addc-563a2eb2efcf service nova] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Refreshing instance network info cache due to event network-changed-c591ec72-086f-44a0-8cf2-3e6719700505. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1083.576426] env[62460]: DEBUG oslo_concurrency.lockutils [req-89538124-ff1f-44fa-8d36-5ab4f0f3bc15 req-0702b7c6-56ee-4df7-addc-563a2eb2efcf service nova] Acquiring lock "refresh_cache-00597ea8-88ee-448d-b5b5-2b29dc8683e9" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1083.576568] env[62460]: DEBUG oslo_concurrency.lockutils [req-89538124-ff1f-44fa-8d36-5ab4f0f3bc15 req-0702b7c6-56ee-4df7-addc-563a2eb2efcf service nova] Acquired lock "refresh_cache-00597ea8-88ee-448d-b5b5-2b29dc8683e9" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.576732] env[62460]: DEBUG nova.network.neutron [req-89538124-ff1f-44fa-8d36-5ab4f0f3bc15 req-0702b7c6-56ee-4df7-addc-563a2eb2efcf service nova] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Refreshing network info cache for port c591ec72-086f-44a0-8cf2-3e6719700505 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1083.604260] env[62460]: DEBUG oslo_concurrency.lockutils [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c6008d-80c8-470b-9f2c-c42ba1856c06" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1083.604507] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Processing image 11c6008d-80c8-470b-9f2c-c42ba1856c06 {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1083.604852] env[62460]: DEBUG oslo_concurrency.lockutils [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c6008d-80c8-470b-9f2c-c42ba1856c06/11c6008d-80c8-470b-9f2c-c42ba1856c06.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1083.604916] env[62460]: DEBUG oslo_concurrency.lockutils [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c6008d-80c8-470b-9f2c-c42ba1856c06/11c6008d-80c8-470b-9f2c-c42ba1856c06.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.605080] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1083.605333] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2d935ff6-c500-4bf7-a2a9-eac6e0f96ec8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.613281] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1083.613504] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1083.614239] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b11f62bc-fb69-4759-9fce-14a33a376794 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.619394] env[62460]: DEBUG oslo_vmware.api [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Waiting for the task: (returnval){ [ 1083.619394] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52257878-d621-710a-e8cb-a27767cbbc1e" [ 1083.619394] env[62460]: _type = "Task" [ 1083.619394] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.626385] env[62460]: DEBUG oslo_vmware.api [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52257878-d621-710a-e8cb-a27767cbbc1e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.708682] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 61073a4e-c5e5-48f5-a43a-ce71bd18e48a] Instance has had 0 of 5 cleanup attempts {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1083.781197] env[62460]: DEBUG nova.scheduler.client.report [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1083.926160] env[62460]: DEBUG oslo_concurrency.lockutils [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquiring lock "00597ea8-88ee-448d-b5b5-2b29dc8683e9" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1084.135940] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Preparing fetch location {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1084.137325] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Fetch image to [datastore1] OSTACK_IMG_39cf137d-7184-4e1b-ae70-40633993c2dd/OSTACK_IMG_39cf137d-7184-4e1b-ae70-40633993c2dd.vmdk {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1084.137610] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Downloading stream optimized image 11c6008d-80c8-470b-9f2c-c42ba1856c06 to [datastore1] OSTACK_IMG_39cf137d-7184-4e1b-ae70-40633993c2dd/OSTACK_IMG_39cf137d-7184-4e1b-ae70-40633993c2dd.vmdk on the data store datastore1 as vApp {{(pid=62460) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1084.137804] env[62460]: DEBUG nova.virt.vmwareapi.images [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Downloading image file data 11c6008d-80c8-470b-9f2c-c42ba1856c06 to the ESX as VM named 'OSTACK_IMG_39cf137d-7184-4e1b-ae70-40633993c2dd' {{(pid=62460) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1084.213362] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: abd4501b-3ad8-4103-921d-5b80fc2f313f] Instance has had 0 of 5 cleanup attempts {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1084.236965] env[62460]: DEBUG oslo_vmware.rw_handles [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1084.236965] env[62460]: value = "resgroup-9" [ 1084.236965] env[62460]: _type = "ResourcePool" [ 1084.236965] env[62460]: }. {{(pid=62460) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1084.237169] env[62460]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-b1d02cd4-1496-416d-9758-5b641aced104 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.262606] env[62460]: DEBUG oslo_vmware.rw_handles [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Lease: (returnval){ [ 1084.262606] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52388eb2-1aad-ead0-c712-63969b610764" [ 1084.262606] env[62460]: _type = "HttpNfcLease" [ 1084.262606] env[62460]: } obtained for vApp import into resource pool (val){ [ 1084.262606] env[62460]: value = "resgroup-9" [ 1084.262606] env[62460]: _type = "ResourcePool" [ 1084.262606] env[62460]: }. {{(pid=62460) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1084.262912] env[62460]: DEBUG oslo_vmware.api [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Waiting for the lease: (returnval){ [ 1084.262912] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52388eb2-1aad-ead0-c712-63969b610764" [ 1084.262912] env[62460]: _type = "HttpNfcLease" [ 1084.262912] env[62460]: } to be ready. {{(pid=62460) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1084.269814] env[62460]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1084.269814] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52388eb2-1aad-ead0-c712-63969b610764" [ 1084.269814] env[62460]: _type = "HttpNfcLease" [ 1084.269814] env[62460]: } is initializing. {{(pid=62460) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1084.286217] env[62460]: DEBUG oslo_concurrency.lockutils [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.125s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1084.333965] env[62460]: DEBUG nova.network.neutron [req-89538124-ff1f-44fa-8d36-5ab4f0f3bc15 req-0702b7c6-56ee-4df7-addc-563a2eb2efcf service nova] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Updated VIF entry in instance network info cache for port c591ec72-086f-44a0-8cf2-3e6719700505. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1084.334367] env[62460]: DEBUG nova.network.neutron [req-89538124-ff1f-44fa-8d36-5ab4f0f3bc15 req-0702b7c6-56ee-4df7-addc-563a2eb2efcf service nova] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Updating instance_info_cache with network_info: [{"id": "c591ec72-086f-44a0-8cf2-3e6719700505", "address": "fa:16:3e:30:d9:88", "network": {"id": "b5cb4ee0-4562-45be-a25d-51c3a5932efd", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-823340443-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.156", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3d2d1c48ec14121a2e8c9b3f800a949", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapc591ec72-08", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1084.718118] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: bf28fc98-d9a9-4678-a107-ace2b6503353] Instance has had 0 of 5 cleanup attempts {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1084.770787] env[62460]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1084.770787] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52388eb2-1aad-ead0-c712-63969b610764" [ 1084.770787] env[62460]: _type = "HttpNfcLease" [ 1084.770787] env[62460]: } is initializing. {{(pid=62460) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1084.795029] env[62460]: DEBUG oslo_concurrency.lockutils [None req-16f89153-ed83-4987-8178-f31066d76578 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lock "00597ea8-88ee-448d-b5b5-2b29dc8683e9" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 20.560s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1084.795414] env[62460]: DEBUG oslo_concurrency.lockutils [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lock "00597ea8-88ee-448d-b5b5-2b29dc8683e9" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.869s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1084.795597] env[62460]: INFO nova.compute.manager [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Unshelving [ 1084.836705] env[62460]: DEBUG oslo_concurrency.lockutils [req-89538124-ff1f-44fa-8d36-5ab4f0f3bc15 req-0702b7c6-56ee-4df7-addc-563a2eb2efcf service nova] Releasing lock "refresh_cache-00597ea8-88ee-448d-b5b5-2b29dc8683e9" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1085.221244] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 6b084ed4-e8f0-4063-bf2a-252740640753] Instance has had 0 of 5 cleanup attempts {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1085.271255] env[62460]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1085.271255] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52388eb2-1aad-ead0-c712-63969b610764" [ 1085.271255] env[62460]: _type = "HttpNfcLease" [ 1085.271255] env[62460]: } is ready. {{(pid=62460) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1085.271596] env[62460]: DEBUG oslo_vmware.rw_handles [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1085.271596] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52388eb2-1aad-ead0-c712-63969b610764" [ 1085.271596] env[62460]: _type = "HttpNfcLease" [ 1085.271596] env[62460]: }. {{(pid=62460) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1085.272250] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c05087d5-9366-4edf-9cba-02c00c8ed2ef {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.279010] env[62460]: DEBUG oslo_vmware.rw_handles [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520a757f-dd2d-79a4-ba3a-ef8c6800831f/disk-0.vmdk from lease info. {{(pid=62460) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1085.279224] env[62460]: DEBUG oslo_vmware.rw_handles [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Creating HTTP connection to write to file with size = 31661056 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520a757f-dd2d-79a4-ba3a-ef8c6800831f/disk-0.vmdk. {{(pid=62460) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1085.343104] env[62460]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-f8271e86-c96e-43f7-ad06-5eb7a815f4ce {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.726413] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 09155629-51c8-4043-b1ee-6d5036552a67] Instance has had 0 of 5 cleanup attempts {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1085.860356] env[62460]: DEBUG oslo_concurrency.lockutils [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1085.860595] env[62460]: DEBUG oslo_concurrency.lockutils [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1085.860869] env[62460]: DEBUG nova.objects.instance [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lazy-loading 'pci_requests' on Instance uuid 00597ea8-88ee-448d-b5b5-2b29dc8683e9 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1086.231731] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: e90e8c3a-d244-4d04-997e-c15ab1cb2009] Instance has had 0 of 5 cleanup attempts {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1086.365823] env[62460]: DEBUG nova.objects.instance [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lazy-loading 'numa_topology' on Instance uuid 00597ea8-88ee-448d-b5b5-2b29dc8683e9 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1086.478354] env[62460]: DEBUG oslo_vmware.rw_handles [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Completed reading data from the image iterator. {{(pid=62460) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1086.478773] env[62460]: DEBUG oslo_vmware.rw_handles [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520a757f-dd2d-79a4-ba3a-ef8c6800831f/disk-0.vmdk. {{(pid=62460) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1086.479861] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d76393f0-e88d-4042-b279-72593ece6b93 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.487393] env[62460]: DEBUG oslo_vmware.rw_handles [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520a757f-dd2d-79a4-ba3a-ef8c6800831f/disk-0.vmdk is in state: ready. {{(pid=62460) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1086.487527] env[62460]: DEBUG oslo_vmware.rw_handles [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520a757f-dd2d-79a4-ba3a-ef8c6800831f/disk-0.vmdk. {{(pid=62460) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1086.487754] env[62460]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-a8565255-0bc4-4461-860a-0ec6ae66fe0c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.736082] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 8bcaa672-97c8-466a-812c-5723e08d0c52] Instance has had 0 of 5 cleanup attempts {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1086.868852] env[62460]: INFO nova.compute.claims [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1087.239844] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: fbe558f1-ce5a-4429-8833-48bdabe91071] Instance has had 0 of 5 cleanup attempts {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1087.248623] env[62460]: DEBUG oslo_vmware.rw_handles [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520a757f-dd2d-79a4-ba3a-ef8c6800831f/disk-0.vmdk. {{(pid=62460) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1087.248623] env[62460]: INFO nova.virt.vmwareapi.images [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Downloaded image file data 11c6008d-80c8-470b-9f2c-c42ba1856c06 [ 1087.249352] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afc15fc5-75eb-429f-ac0b-61617112d994 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.265609] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-35cfc8d9-96c7-4e7d-826b-041e5482713a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.376561] env[62460]: INFO nova.virt.vmwareapi.images [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] The imported VM was unregistered [ 1087.378904] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Caching image {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1087.379162] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Creating directory with path [datastore1] devstack-image-cache_base/11c6008d-80c8-470b-9f2c-c42ba1856c06 {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1087.379430] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-740eab63-9302-4787-9502-9aebd9d9ac70 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.398833] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Created directory with path [datastore1] devstack-image-cache_base/11c6008d-80c8-470b-9f2c-c42ba1856c06 {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1087.399037] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_39cf137d-7184-4e1b-ae70-40633993c2dd/OSTACK_IMG_39cf137d-7184-4e1b-ae70-40633993c2dd.vmdk to [datastore1] devstack-image-cache_base/11c6008d-80c8-470b-9f2c-c42ba1856c06/11c6008d-80c8-470b-9f2c-c42ba1856c06.vmdk. {{(pid=62460) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1087.399288] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-29b93047-8542-47ca-9a0f-d1b472f0ba21 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.407216] env[62460]: DEBUG oslo_vmware.api [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Waiting for the task: (returnval){ [ 1087.407216] env[62460]: value = "task-1314122" [ 1087.407216] env[62460]: _type = "Task" [ 1087.407216] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.415094] env[62460]: DEBUG oslo_vmware.api [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314122, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.744163] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 0da70720-197c-4483-93e1-019e0f040c4c] Instance has had 0 of 5 cleanup attempts {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1087.917856] env[62460]: DEBUG oslo_vmware.api [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314122, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.931953] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-030c712f-1447-4d76-95fa-c38a944eef44 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.940110] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52fabe7a-9c69-4f9e-bc53-322719026e3d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.971793] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3f9b95c-dacb-4070-9c9f-a1f79920604b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.979507] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56aebcfc-13d5-48e7-84c7-72501624bb9f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.993427] env[62460]: DEBUG nova.compute.provider_tree [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1088.248300] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: fd03509c-5471-48bb-8150-d29531dca848] Instance has had 0 of 5 cleanup attempts {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1088.420750] env[62460]: DEBUG oslo_vmware.api [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314122, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.497126] env[62460]: DEBUG nova.scheduler.client.report [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1088.751986] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 3e10af70-db52-4d4a-bb92-821a05dcbab6] Instance has had 0 of 5 cleanup attempts {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1088.920151] env[62460]: DEBUG oslo_vmware.api [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314122, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.002471] env[62460]: DEBUG oslo_concurrency.lockutils [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.142s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1089.048191] env[62460]: INFO nova.network.neutron [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Updating port c591ec72-086f-44a0-8cf2-3e6719700505 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1089.255518] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 0f5457a0-c918-4407-9ba9-bfeeb949c6a7] Instance has had 0 of 5 cleanup attempts {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1089.421680] env[62460]: DEBUG oslo_vmware.api [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314122, 'name': MoveVirtualDisk_Task} progress is 83%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.759469] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: f1921907-f69c-49a3-896c-a4b12b526ffa] Instance has had 0 of 5 cleanup attempts {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1089.920703] env[62460]: DEBUG oslo_vmware.api [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314122, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.263566] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 896f8ff4-e45f-4403-a727-03ee25e58609] Instance has had 0 of 5 cleanup attempts {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1090.422046] env[62460]: DEBUG oslo_vmware.api [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314122, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.570566} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.422385] env[62460]: INFO nova.virt.vmwareapi.ds_util [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_39cf137d-7184-4e1b-ae70-40633993c2dd/OSTACK_IMG_39cf137d-7184-4e1b-ae70-40633993c2dd.vmdk to [datastore1] devstack-image-cache_base/11c6008d-80c8-470b-9f2c-c42ba1856c06/11c6008d-80c8-470b-9f2c-c42ba1856c06.vmdk. [ 1090.422763] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Cleaning up location [datastore1] OSTACK_IMG_39cf137d-7184-4e1b-ae70-40633993c2dd {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1090.422839] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_39cf137d-7184-4e1b-ae70-40633993c2dd {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1090.423072] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cdc3efe3-6e19-41c5-a070-66f08240884d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.428970] env[62460]: DEBUG oslo_vmware.api [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Waiting for the task: (returnval){ [ 1090.428970] env[62460]: value = "task-1314123" [ 1090.428970] env[62460]: _type = "Task" [ 1090.428970] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.436436] env[62460]: DEBUG oslo_vmware.api [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314123, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.497775] env[62460]: DEBUG nova.compute.manager [req-4f2094ed-1c68-49f5-935e-676736934201 req-b6cad50a-f0fc-457a-b1e6-51d2b1fa5f2f service nova] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Received event network-vif-plugged-c591ec72-086f-44a0-8cf2-3e6719700505 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1090.498018] env[62460]: DEBUG oslo_concurrency.lockutils [req-4f2094ed-1c68-49f5-935e-676736934201 req-b6cad50a-f0fc-457a-b1e6-51d2b1fa5f2f service nova] Acquiring lock "00597ea8-88ee-448d-b5b5-2b29dc8683e9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1090.498254] env[62460]: DEBUG oslo_concurrency.lockutils [req-4f2094ed-1c68-49f5-935e-676736934201 req-b6cad50a-f0fc-457a-b1e6-51d2b1fa5f2f service nova] Lock "00597ea8-88ee-448d-b5b5-2b29dc8683e9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1090.498439] env[62460]: DEBUG oslo_concurrency.lockutils [req-4f2094ed-1c68-49f5-935e-676736934201 req-b6cad50a-f0fc-457a-b1e6-51d2b1fa5f2f service nova] Lock "00597ea8-88ee-448d-b5b5-2b29dc8683e9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.498616] env[62460]: DEBUG nova.compute.manager [req-4f2094ed-1c68-49f5-935e-676736934201 req-b6cad50a-f0fc-457a-b1e6-51d2b1fa5f2f service nova] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] No waiting events found dispatching network-vif-plugged-c591ec72-086f-44a0-8cf2-3e6719700505 {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1090.498809] env[62460]: WARNING nova.compute.manager [req-4f2094ed-1c68-49f5-935e-676736934201 req-b6cad50a-f0fc-457a-b1e6-51d2b1fa5f2f service nova] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Received unexpected event network-vif-plugged-c591ec72-086f-44a0-8cf2-3e6719700505 for instance with vm_state shelved_offloaded and task_state spawning. [ 1090.603599] env[62460]: DEBUG oslo_concurrency.lockutils [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquiring lock "refresh_cache-00597ea8-88ee-448d-b5b5-2b29dc8683e9" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1090.603796] env[62460]: DEBUG oslo_concurrency.lockutils [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquired lock "refresh_cache-00597ea8-88ee-448d-b5b5-2b29dc8683e9" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1090.603977] env[62460]: DEBUG nova.network.neutron [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1090.766801] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 49477d35-92ea-4f9d-8333-fc23144b7dfc] Instance has had 0 of 5 cleanup attempts {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1090.938866] env[62460]: DEBUG oslo_vmware.api [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314123, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.052141} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.939127] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1090.939257] env[62460]: DEBUG oslo_concurrency.lockutils [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c6008d-80c8-470b-9f2c-c42ba1856c06/11c6008d-80c8-470b-9f2c-c42ba1856c06.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1090.939509] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11c6008d-80c8-470b-9f2c-c42ba1856c06/11c6008d-80c8-470b-9f2c-c42ba1856c06.vmdk to [datastore1] c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac/c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1090.939759] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-128ae1ce-57b0-4f49-8f00-2f1ee6b5f46c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.945921] env[62460]: DEBUG oslo_vmware.api [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Waiting for the task: (returnval){ [ 1090.945921] env[62460]: value = "task-1314124" [ 1090.945921] env[62460]: _type = "Task" [ 1090.945921] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.954098] env[62460]: DEBUG oslo_vmware.api [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314124, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.270616] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: a02b4be1-91b0-4254-8d60-654885e24f6b] Instance has had 0 of 5 cleanup attempts {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1091.335413] env[62460]: DEBUG nova.network.neutron [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Updating instance_info_cache with network_info: [{"id": "c591ec72-086f-44a0-8cf2-3e6719700505", "address": "fa:16:3e:30:d9:88", "network": {"id": "b5cb4ee0-4562-45be-a25d-51c3a5932efd", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-823340443-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.156", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3d2d1c48ec14121a2e8c9b3f800a949", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55c757ac-f8b2-466d-b634-07dbd100b312", "external-id": "nsx-vlan-transportzone-159", "segmentation_id": 159, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc591ec72-08", "ovs_interfaceid": "c591ec72-086f-44a0-8cf2-3e6719700505", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1091.456761] env[62460]: DEBUG oslo_vmware.api [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314124, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.774637] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 16d40829-ba6e-4193-98b5-fff94c066bcb] Instance has had 0 of 5 cleanup attempts {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1091.841407] env[62460]: DEBUG oslo_concurrency.lockutils [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Releasing lock "refresh_cache-00597ea8-88ee-448d-b5b5-2b29dc8683e9" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1091.870615] env[62460]: DEBUG nova.virt.hardware [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='423f1f8241ff43e85df13c00a92c6c06',container_format='bare',created_at=2024-10-18T09:09:06Z,direct_url=,disk_format='vmdk',id=5b45a9b7-790b-42da-bcad-cc40f09629fb,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-982209263-shelved',owner='f3d2d1c48ec14121a2e8c9b3f800a949',properties=ImageMetaProps,protected=,size=31667712,status='active',tags=,updated_at=2024-10-18T09:09:20Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1091.870898] env[62460]: DEBUG nova.virt.hardware [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1091.871077] env[62460]: DEBUG nova.virt.hardware [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1091.871275] env[62460]: DEBUG nova.virt.hardware [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1091.871485] env[62460]: DEBUG nova.virt.hardware [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1091.871646] env[62460]: DEBUG nova.virt.hardware [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1091.871864] env[62460]: DEBUG nova.virt.hardware [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1091.872044] env[62460]: DEBUG nova.virt.hardware [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1091.872231] env[62460]: DEBUG nova.virt.hardware [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1091.872410] env[62460]: DEBUG nova.virt.hardware [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1091.872621] env[62460]: DEBUG nova.virt.hardware [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1091.873527] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d6f4ee4-6670-418b-a106-a8ca9d29f13b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.881786] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b3ce621-a0f1-4310-987b-1d460fce8d09 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.895690] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:30:d9:88', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '55c757ac-f8b2-466d-b634-07dbd100b312', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c591ec72-086f-44a0-8cf2-3e6719700505', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1091.902980] env[62460]: DEBUG oslo.service.loopingcall [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1091.903252] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1091.903482] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b8802497-8706-446a-9683-cac571db6c7c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.922942] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1091.922942] env[62460]: value = "task-1314125" [ 1091.922942] env[62460]: _type = "Task" [ 1091.922942] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.931384] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1314125, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.957481] env[62460]: DEBUG oslo_vmware.api [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314124, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.278539] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 9d1a5830-f3c0-4d18-9338-16f7b6962c6a] Instance has had 0 of 5 cleanup attempts {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1092.433271] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1314125, 'name': CreateVM_Task} progress is 99%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.457151] env[62460]: DEBUG oslo_vmware.api [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314124, 'name': CopyVirtualDisk_Task} progress is 60%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.528231] env[62460]: DEBUG nova.compute.manager [req-707db341-cc22-4fb7-958e-da943bf00ef9 req-78a83c58-7b4b-4ff0-a8c0-8555ae62eaa0 service nova] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Received event network-changed-c591ec72-086f-44a0-8cf2-3e6719700505 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1092.528501] env[62460]: DEBUG nova.compute.manager [req-707db341-cc22-4fb7-958e-da943bf00ef9 req-78a83c58-7b4b-4ff0-a8c0-8555ae62eaa0 service nova] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Refreshing instance network info cache due to event network-changed-c591ec72-086f-44a0-8cf2-3e6719700505. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1092.528787] env[62460]: DEBUG oslo_concurrency.lockutils [req-707db341-cc22-4fb7-958e-da943bf00ef9 req-78a83c58-7b4b-4ff0-a8c0-8555ae62eaa0 service nova] Acquiring lock "refresh_cache-00597ea8-88ee-448d-b5b5-2b29dc8683e9" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1092.528951] env[62460]: DEBUG oslo_concurrency.lockutils [req-707db341-cc22-4fb7-958e-da943bf00ef9 req-78a83c58-7b4b-4ff0-a8c0-8555ae62eaa0 service nova] Acquired lock "refresh_cache-00597ea8-88ee-448d-b5b5-2b29dc8683e9" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1092.529187] env[62460]: DEBUG nova.network.neutron [req-707db341-cc22-4fb7-958e-da943bf00ef9 req-78a83c58-7b4b-4ff0-a8c0-8555ae62eaa0 service nova] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Refreshing network info cache for port c591ec72-086f-44a0-8cf2-3e6719700505 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1092.783212] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 0269dc64-d2b1-43c5-bdf7-11d97e534819] Instance has had 0 of 5 cleanup attempts {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1092.934535] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1314125, 'name': CreateVM_Task, 'duration_secs': 0.606491} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.934766] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1092.935480] env[62460]: DEBUG oslo_concurrency.lockutils [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5b45a9b7-790b-42da-bcad-cc40f09629fb" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1092.935656] env[62460]: DEBUG oslo_concurrency.lockutils [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5b45a9b7-790b-42da-bcad-cc40f09629fb" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1092.936112] env[62460]: DEBUG oslo_concurrency.lockutils [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/5b45a9b7-790b-42da-bcad-cc40f09629fb" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1092.936402] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9424ac04-ef31-462d-970c-3736f6a6cfa4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.941447] env[62460]: DEBUG oslo_vmware.api [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 1092.941447] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52999cbf-8659-18fa-8fd6-e068b60d2c89" [ 1092.941447] env[62460]: _type = "Task" [ 1092.941447] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.951377] env[62460]: DEBUG oslo_vmware.api [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52999cbf-8659-18fa-8fd6-e068b60d2c89, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.959786] env[62460]: DEBUG oslo_vmware.api [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314124, 'name': CopyVirtualDisk_Task} progress is 80%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.283214] env[62460]: DEBUG nova.network.neutron [req-707db341-cc22-4fb7-958e-da943bf00ef9 req-78a83c58-7b4b-4ff0-a8c0-8555ae62eaa0 service nova] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Updated VIF entry in instance network info cache for port c591ec72-086f-44a0-8cf2-3e6719700505. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1093.283624] env[62460]: DEBUG nova.network.neutron [req-707db341-cc22-4fb7-958e-da943bf00ef9 req-78a83c58-7b4b-4ff0-a8c0-8555ae62eaa0 service nova] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Updating instance_info_cache with network_info: [{"id": "c591ec72-086f-44a0-8cf2-3e6719700505", "address": "fa:16:3e:30:d9:88", "network": {"id": "b5cb4ee0-4562-45be-a25d-51c3a5932efd", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-823340443-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.156", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3d2d1c48ec14121a2e8c9b3f800a949", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55c757ac-f8b2-466d-b634-07dbd100b312", "external-id": "nsx-vlan-transportzone-159", "segmentation_id": 159, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc591ec72-08", "ovs_interfaceid": "c591ec72-086f-44a0-8cf2-3e6719700505", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1093.286827] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 9f81c6e9-1e73-4e6d-be6f-3b88d1243a0b] Instance has had 0 of 5 cleanup attempts {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1093.454949] env[62460]: DEBUG oslo_concurrency.lockutils [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5b45a9b7-790b-42da-bcad-cc40f09629fb" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1093.455189] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Processing image 5b45a9b7-790b-42da-bcad-cc40f09629fb {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1093.455448] env[62460]: DEBUG oslo_concurrency.lockutils [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5b45a9b7-790b-42da-bcad-cc40f09629fb/5b45a9b7-790b-42da-bcad-cc40f09629fb.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1093.455600] env[62460]: DEBUG oslo_concurrency.lockutils [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5b45a9b7-790b-42da-bcad-cc40f09629fb/5b45a9b7-790b-42da-bcad-cc40f09629fb.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1093.455788] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1093.456276] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-374c5871-f987-45cf-8292-b74863146c5d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.462732] env[62460]: DEBUG oslo_vmware.api [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314124, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.477053] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1093.477251] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1093.478071] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3643972a-ae19-47f7-aedd-192dcc12274e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.483156] env[62460]: DEBUG oslo_vmware.api [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 1093.483156] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]5239e6ad-5bad-4c2f-0da4-d90fa9cc1db4" [ 1093.483156] env[62460]: _type = "Task" [ 1093.483156] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.490695] env[62460]: DEBUG oslo_vmware.api [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5239e6ad-5bad-4c2f-0da4-d90fa9cc1db4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.786825] env[62460]: DEBUG oslo_concurrency.lockutils [req-707db341-cc22-4fb7-958e-da943bf00ef9 req-78a83c58-7b4b-4ff0-a8c0-8555ae62eaa0 service nova] Releasing lock "refresh_cache-00597ea8-88ee-448d-b5b5-2b29dc8683e9" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1093.789551] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 5b6e8205-003e-49c0-a73d-be2e032a8272] Instance has had 0 of 5 cleanup attempts {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1093.963746] env[62460]: DEBUG oslo_vmware.api [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314124, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.6215} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.963746] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11c6008d-80c8-470b-9f2c-c42ba1856c06/11c6008d-80c8-470b-9f2c-c42ba1856c06.vmdk to [datastore1] c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac/c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1093.963959] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c17f6ed-5111-4181-9517-a1becf455328 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.985013] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Reconfiguring VM instance instance-00000066 to attach disk [datastore1] c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac/c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac.vmdk or device None with type streamOptimized {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1093.985316] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ae218ca8-761b-4431-b54d-8f538bcbfb17 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.008252] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Preparing fetch location {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1094.008512] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Fetch image to [datastore2] OSTACK_IMG_5baa665f-b874-4d49-b10e-e2a99f2e66e8/OSTACK_IMG_5baa665f-b874-4d49-b10e-e2a99f2e66e8.vmdk {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1094.008715] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Downloading stream optimized image 5b45a9b7-790b-42da-bcad-cc40f09629fb to [datastore2] OSTACK_IMG_5baa665f-b874-4d49-b10e-e2a99f2e66e8/OSTACK_IMG_5baa665f-b874-4d49-b10e-e2a99f2e66e8.vmdk on the data store datastore2 as vApp {{(pid=62460) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1094.008925] env[62460]: DEBUG nova.virt.vmwareapi.images [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Downloading image file data 5b45a9b7-790b-42da-bcad-cc40f09629fb to the ESX as VM named 'OSTACK_IMG_5baa665f-b874-4d49-b10e-e2a99f2e66e8' {{(pid=62460) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1094.010884] env[62460]: DEBUG oslo_vmware.api [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Waiting for the task: (returnval){ [ 1094.010884] env[62460]: value = "task-1314126" [ 1094.010884] env[62460]: _type = "Task" [ 1094.010884] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.018285] env[62460]: DEBUG oslo_vmware.api [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314126, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.077272] env[62460]: DEBUG oslo_vmware.rw_handles [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1094.077272] env[62460]: value = "resgroup-9" [ 1094.077272] env[62460]: _type = "ResourcePool" [ 1094.077272] env[62460]: }. {{(pid=62460) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1094.077640] env[62460]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-6763e711-a177-44ff-b70c-557a38d3c953 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.098019] env[62460]: DEBUG oslo_vmware.rw_handles [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lease: (returnval){ [ 1094.098019] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52f07de5-8da6-9c36-22c4-be3b3e46dda3" [ 1094.098019] env[62460]: _type = "HttpNfcLease" [ 1094.098019] env[62460]: } obtained for vApp import into resource pool (val){ [ 1094.098019] env[62460]: value = "resgroup-9" [ 1094.098019] env[62460]: _type = "ResourcePool" [ 1094.098019] env[62460]: }. {{(pid=62460) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1094.098335] env[62460]: DEBUG oslo_vmware.api [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the lease: (returnval){ [ 1094.098335] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52f07de5-8da6-9c36-22c4-be3b3e46dda3" [ 1094.098335] env[62460]: _type = "HttpNfcLease" [ 1094.098335] env[62460]: } to be ready. {{(pid=62460) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1094.105924] env[62460]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1094.105924] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52f07de5-8da6-9c36-22c4-be3b3e46dda3" [ 1094.105924] env[62460]: _type = "HttpNfcLease" [ 1094.105924] env[62460]: } is initializing. {{(pid=62460) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1094.293389] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: b76f37a0-91d0-4a01-9d95-9c6586081175] Instance has had 0 of 5 cleanup attempts {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1094.520902] env[62460]: DEBUG oslo_vmware.api [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314126, 'name': ReconfigVM_Task, 'duration_secs': 0.371281} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.521195] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Reconfigured VM instance instance-00000066 to attach disk [datastore1] c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac/c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac.vmdk or device None with type streamOptimized {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1094.521862] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-39da0ba7-ef13-4f44-81fd-cd4d2642b2cf {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.527817] env[62460]: DEBUG oslo_vmware.api [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Waiting for the task: (returnval){ [ 1094.527817] env[62460]: value = "task-1314128" [ 1094.527817] env[62460]: _type = "Task" [ 1094.527817] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.536469] env[62460]: DEBUG oslo_vmware.api [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314128, 'name': Rename_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.605487] env[62460]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1094.605487] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52f07de5-8da6-9c36-22c4-be3b3e46dda3" [ 1094.605487] env[62460]: _type = "HttpNfcLease" [ 1094.605487] env[62460]: } is initializing. {{(pid=62460) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1094.797305] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 13890351-6091-4b4f-8484-1cdd0c8523b0] Instance has had 0 of 5 cleanup attempts {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1095.038464] env[62460]: DEBUG oslo_vmware.api [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314128, 'name': Rename_Task, 'duration_secs': 0.162585} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.038786] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1095.039059] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ea377108-4899-46c7-9e2b-b7ceb492c257 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.045030] env[62460]: DEBUG oslo_vmware.api [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Waiting for the task: (returnval){ [ 1095.045030] env[62460]: value = "task-1314129" [ 1095.045030] env[62460]: _type = "Task" [ 1095.045030] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.052495] env[62460]: DEBUG oslo_vmware.api [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314129, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.106103] env[62460]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1095.106103] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52f07de5-8da6-9c36-22c4-be3b3e46dda3" [ 1095.106103] env[62460]: _type = "HttpNfcLease" [ 1095.106103] env[62460]: } is ready. {{(pid=62460) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1095.106525] env[62460]: DEBUG oslo_vmware.rw_handles [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1095.106525] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52f07de5-8da6-9c36-22c4-be3b3e46dda3" [ 1095.106525] env[62460]: _type = "HttpNfcLease" [ 1095.106525] env[62460]: }. {{(pid=62460) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1095.107353] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5ad7afb-b6fd-413c-bddf-34bbede2028e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.114286] env[62460]: DEBUG oslo_vmware.rw_handles [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526daaed-7215-ef36-f0a4-c213e378a41d/disk-0.vmdk from lease info. {{(pid=62460) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1095.114590] env[62460]: DEBUG oslo_vmware.rw_handles [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Creating HTTP connection to write to file with size = 31667712 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526daaed-7215-ef36-f0a4-c213e378a41d/disk-0.vmdk. {{(pid=62460) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1095.179257] env[62460]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-efa5c024-f8a6-446b-9c03-3684440b25ff {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.300636] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: fde12685-d3b9-46a0-8931-25b904d4f21e] Instance has had 0 of 5 cleanup attempts {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1095.557822] env[62460]: DEBUG oslo_vmware.api [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314129, 'name': PowerOnVM_Task, 'duration_secs': 0.463832} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.560842] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1095.667374] env[62460]: DEBUG nova.compute.manager [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1095.668345] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-846686d0-dbca-4f66-b0c0-b3dac9092805 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.804633] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 3ada3516-3147-4566-a46a-1cb29cf880d0] Instance has had 0 of 5 cleanup attempts {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1096.187114] env[62460]: DEBUG oslo_concurrency.lockutils [None req-04226ef9-fedb-4805-802b-97db6b9c0e71 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Lock "c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 21.654s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.308533] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 72e91f8e-0619-464c-b9bc-d6a14be42cb8] Instance has had 0 of 5 cleanup attempts {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1096.312632] env[62460]: DEBUG oslo_vmware.rw_handles [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Completed reading data from the image iterator. {{(pid=62460) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1096.312930] env[62460]: DEBUG oslo_vmware.rw_handles [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526daaed-7215-ef36-f0a4-c213e378a41d/disk-0.vmdk. {{(pid=62460) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1096.313873] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-828cd948-d978-425b-b608-85fa00f8c5b5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.320595] env[62460]: DEBUG oslo_vmware.rw_handles [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526daaed-7215-ef36-f0a4-c213e378a41d/disk-0.vmdk is in state: ready. {{(pid=62460) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1096.321231] env[62460]: DEBUG oslo_vmware.rw_handles [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526daaed-7215-ef36-f0a4-c213e378a41d/disk-0.vmdk. {{(pid=62460) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1096.321231] env[62460]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-c1cd23a0-64cd-468b-b4d9-7318faa3bc06 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.817641] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: db09d1f5-88cc-4dc7-9a7b-5d53d09567fa] Instance has had 0 of 5 cleanup attempts {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1096.918640] env[62460]: DEBUG oslo_vmware.rw_handles [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526daaed-7215-ef36-f0a4-c213e378a41d/disk-0.vmdk. {{(pid=62460) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1096.918928] env[62460]: INFO nova.virt.vmwareapi.images [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Downloaded image file data 5b45a9b7-790b-42da-bcad-cc40f09629fb [ 1096.919811] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3375579d-9754-46d2-a4f7-4f92952a6871 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.934807] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4d8c7620-be56-402d-a397-120ea6dbdddc {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.066261] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8c7e5df-4a8c-4812-bcbb-12b8a1ad3947 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.072888] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-b333af41-6627-4123-a51b-966b836bc548 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Suspending the VM {{(pid=62460) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1097.073145] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-26cdff1c-975a-4989-a6f6-61bce3a9b983 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.079618] env[62460]: DEBUG oslo_vmware.api [None req-b333af41-6627-4123-a51b-966b836bc548 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Waiting for the task: (returnval){ [ 1097.079618] env[62460]: value = "task-1314131" [ 1097.079618] env[62460]: _type = "Task" [ 1097.079618] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.086716] env[62460]: DEBUG oslo_vmware.api [None req-b333af41-6627-4123-a51b-966b836bc548 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314131, 'name': SuspendVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.103945] env[62460]: INFO nova.virt.vmwareapi.images [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] The imported VM was unregistered [ 1097.106560] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Caching image {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1097.106823] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Creating directory with path [datastore2] devstack-image-cache_base/5b45a9b7-790b-42da-bcad-cc40f09629fb {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1097.107112] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-792572b4-43dd-4c60-87c2-f92dc20a0ccc {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.118314] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Created directory with path [datastore2] devstack-image-cache_base/5b45a9b7-790b-42da-bcad-cc40f09629fb {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1097.118528] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_5baa665f-b874-4d49-b10e-e2a99f2e66e8/OSTACK_IMG_5baa665f-b874-4d49-b10e-e2a99f2e66e8.vmdk to [datastore2] devstack-image-cache_base/5b45a9b7-790b-42da-bcad-cc40f09629fb/5b45a9b7-790b-42da-bcad-cc40f09629fb.vmdk. {{(pid=62460) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1097.118802] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-ce6b197b-e419-42d4-bdab-6c10a5ce996a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.125685] env[62460]: DEBUG oslo_vmware.api [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 1097.125685] env[62460]: value = "task-1314132" [ 1097.125685] env[62460]: _type = "Task" [ 1097.125685] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.133585] env[62460]: DEBUG oslo_vmware.api [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314132, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.321501] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 7e6ff902-4a04-43d5-9014-38c4ec88efc4] Instance has had 0 of 5 cleanup attempts {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1097.591358] env[62460]: DEBUG oslo_vmware.api [None req-b333af41-6627-4123-a51b-966b836bc548 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314131, 'name': SuspendVM_Task} progress is 62%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.636077] env[62460]: DEBUG oslo_vmware.api [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314132, 'name': MoveVirtualDisk_Task} progress is 15%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.825023] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 1f318a64-2c38-470b-8fae-4ba4543a5681] Instance has had 0 of 5 cleanup attempts {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1098.090491] env[62460]: DEBUG oslo_vmware.api [None req-b333af41-6627-4123-a51b-966b836bc548 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314131, 'name': SuspendVM_Task, 'duration_secs': 0.761733} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.090795] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-b333af41-6627-4123-a51b-966b836bc548 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Suspended the VM {{(pid=62460) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1098.091028] env[62460]: DEBUG nova.compute.manager [None req-b333af41-6627-4123-a51b-966b836bc548 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1098.091851] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64a81c3a-3a1e-4182-a9ed-611c6a2ba5c3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.137808] env[62460]: DEBUG oslo_vmware.api [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314132, 'name': MoveVirtualDisk_Task} progress is 38%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.328222] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 51718896-f5bf-43a9-9396-1ac768737ba2] Instance has had 0 of 5 cleanup attempts {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1098.637589] env[62460]: DEBUG oslo_vmware.api [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314132, 'name': MoveVirtualDisk_Task} progress is 60%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.831767] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 4ed90d16-81a6-4dbd-8936-0e137151171f] Instance has had 0 of 5 cleanup attempts {{(pid=62460) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1099.138111] env[62460]: DEBUG oslo_vmware.api [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314132, 'name': MoveVirtualDisk_Task} progress is 83%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.335179] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1099.447692] env[62460]: INFO nova.compute.manager [None req-e10c6ab1-73a8-41d7-bbe8-8c1344df9372 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Resuming [ 1099.448402] env[62460]: DEBUG nova.objects.instance [None req-e10c6ab1-73a8-41d7-bbe8-8c1344df9372 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Lazy-loading 'flavor' on Instance uuid c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1099.638940] env[62460]: DEBUG oslo_vmware.api [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314132, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.140111] env[62460]: DEBUG oslo_vmware.api [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314132, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.714174} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.140405] env[62460]: INFO nova.virt.vmwareapi.ds_util [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_5baa665f-b874-4d49-b10e-e2a99f2e66e8/OSTACK_IMG_5baa665f-b874-4d49-b10e-e2a99f2e66e8.vmdk to [datastore2] devstack-image-cache_base/5b45a9b7-790b-42da-bcad-cc40f09629fb/5b45a9b7-790b-42da-bcad-cc40f09629fb.vmdk. [ 1100.140661] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Cleaning up location [datastore2] OSTACK_IMG_5baa665f-b874-4d49-b10e-e2a99f2e66e8 {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1100.140874] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_5baa665f-b874-4d49-b10e-e2a99f2e66e8 {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1100.141478] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-356967c7-34b6-4fdf-bd56-4303d1c24d4a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.146924] env[62460]: DEBUG oslo_vmware.api [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 1100.146924] env[62460]: value = "task-1314133" [ 1100.146924] env[62460]: _type = "Task" [ 1100.146924] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.154285] env[62460]: DEBUG oslo_vmware.api [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314133, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.457280] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e10c6ab1-73a8-41d7-bbe8-8c1344df9372 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Acquiring lock "refresh_cache-c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1100.457446] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e10c6ab1-73a8-41d7-bbe8-8c1344df9372 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Acquired lock "refresh_cache-c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1100.457590] env[62460]: DEBUG nova.network.neutron [None req-e10c6ab1-73a8-41d7-bbe8-8c1344df9372 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1100.657013] env[62460]: DEBUG oslo_vmware.api [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314133, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.0381} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.657297] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1100.657473] env[62460]: DEBUG oslo_concurrency.lockutils [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5b45a9b7-790b-42da-bcad-cc40f09629fb/5b45a9b7-790b-42da-bcad-cc40f09629fb.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1100.657726] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/5b45a9b7-790b-42da-bcad-cc40f09629fb/5b45a9b7-790b-42da-bcad-cc40f09629fb.vmdk to [datastore2] 00597ea8-88ee-448d-b5b5-2b29dc8683e9/00597ea8-88ee-448d-b5b5-2b29dc8683e9.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1100.657982] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d4691b9d-63f9-40e2-8eaf-51caccfee4b2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.664598] env[62460]: DEBUG oslo_vmware.api [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 1100.664598] env[62460]: value = "task-1314134" [ 1100.664598] env[62460]: _type = "Task" [ 1100.664598] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.672091] env[62460]: DEBUG oslo_vmware.api [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314134, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.170203] env[62460]: DEBUG nova.network.neutron [None req-e10c6ab1-73a8-41d7-bbe8-8c1344df9372 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Updating instance_info_cache with network_info: [{"id": "d15647d8-f541-4169-bea9-0937db3301b1", "address": "fa:16:3e:28:b4:d2", "network": {"id": "3aa5e026-c4ee-4641-831b-a7e31b69adc7", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-702747461-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f13597d128c44e19b7f92f5ba1ba217", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b91b49a8-b849-4d0c-97f7-74fdcd88ae03", "external-id": "nsx-vlan-transportzone-406", "segmentation_id": 406, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd15647d8-f5", "ovs_interfaceid": "d15647d8-f541-4169-bea9-0937db3301b1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1101.176864] env[62460]: DEBUG oslo_vmware.api [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314134, 'name': CopyVirtualDisk_Task} progress is 18%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.324375] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1101.324615] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1101.676383] env[62460]: DEBUG oslo_concurrency.lockutils [None req-e10c6ab1-73a8-41d7-bbe8-8c1344df9372 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Releasing lock "refresh_cache-c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1101.676845] env[62460]: DEBUG oslo_vmware.api [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314134, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.677737] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-292da6aa-0348-4078-a5ad-fdd8041018bc {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.684370] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-e10c6ab1-73a8-41d7-bbe8-8c1344df9372 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Resuming the VM {{(pid=62460) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 1101.684664] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4a952991-caed-42e4-8bf8-43d63ce4576d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.691032] env[62460]: DEBUG oslo_vmware.api [None req-e10c6ab1-73a8-41d7-bbe8-8c1344df9372 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Waiting for the task: (returnval){ [ 1101.691032] env[62460]: value = "task-1314135" [ 1101.691032] env[62460]: _type = "Task" [ 1101.691032] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.699629] env[62460]: DEBUG oslo_vmware.api [None req-e10c6ab1-73a8-41d7-bbe8-8c1344df9372 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314135, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.830076] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._sync_power_states {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1102.175900] env[62460]: DEBUG oslo_vmware.api [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314134, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.201029] env[62460]: DEBUG oslo_vmware.api [None req-e10c6ab1-73a8-41d7-bbe8-8c1344df9372 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314135, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.333679] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Getting list of instances from cluster (obj){ [ 1102.333679] env[62460]: value = "domain-c8" [ 1102.333679] env[62460]: _type = "ClusterComputeResource" [ 1102.333679] env[62460]: } {{(pid=62460) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 1102.334785] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30a36ee0-4e8c-440f-b2e5-16d94797d23c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.345794] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Got total of 3 instances {{(pid=62460) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 1102.346128] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Triggering sync for uuid 61f4666e-0588-4a00-a969-2349148caef2 {{(pid=62460) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1102.346186] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Triggering sync for uuid c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac {{(pid=62460) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1102.346350] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Triggering sync for uuid 00597ea8-88ee-448d-b5b5-2b29dc8683e9 {{(pid=62460) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1102.346731] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Acquiring lock "61f4666e-0588-4a00-a969-2349148caef2" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.346977] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "61f4666e-0588-4a00-a969-2349148caef2" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.347360] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Acquiring lock "c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.347569] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.347786] env[62460]: INFO nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] During sync_power_state the instance has a pending task (resuming). Skip. [ 1102.347972] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1102.348193] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Acquiring lock "00597ea8-88ee-448d-b5b5-2b29dc8683e9" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.348391] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1102.349950] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34a24809-6ddc-4e0f-beb8-77c4ef568e60 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.675781] env[62460]: DEBUG oslo_vmware.api [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314134, 'name': CopyVirtualDisk_Task} progress is 85%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.701998] env[62460]: DEBUG oslo_vmware.api [None req-e10c6ab1-73a8-41d7-bbe8-8c1344df9372 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314135, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.854478] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.854783] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.855035] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1102.855262] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62460) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1102.856243] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7be94545-f1da-479e-9438-8797e5bfae5f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.859476] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "61f4666e-0588-4a00-a969-2349148caef2" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.512s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1102.865331] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-736bb7c2-715b-4641-a57e-188d72c82d9b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.880807] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2aa2d3a-69ed-4f7f-960d-93836b0b7919 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.888682] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e584790-85da-4714-81d8-d040f5b04cc5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.919798] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180916MB free_disk=135GB free_vcpus=48 pci_devices=None {{(pid=62460) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1102.919972] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.920194] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1103.177126] env[62460]: DEBUG oslo_vmware.api [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314134, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.300504} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.177442] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/5b45a9b7-790b-42da-bcad-cc40f09629fb/5b45a9b7-790b-42da-bcad-cc40f09629fb.vmdk to [datastore2] 00597ea8-88ee-448d-b5b5-2b29dc8683e9/00597ea8-88ee-448d-b5b5-2b29dc8683e9.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1103.178241] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26165517-f782-4a1a-aba4-f52ec50e0805 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.200386] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] 00597ea8-88ee-448d-b5b5-2b29dc8683e9/00597ea8-88ee-448d-b5b5-2b29dc8683e9.vmdk or device None with type streamOptimized {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1103.201077] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-18f3bde9-a676-47bf-9b76-f25323357ceb {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.223453] env[62460]: DEBUG oslo_vmware.api [None req-e10c6ab1-73a8-41d7-bbe8-8c1344df9372 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314135, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.224874] env[62460]: DEBUG oslo_vmware.api [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 1103.224874] env[62460]: value = "task-1314136" [ 1103.224874] env[62460]: _type = "Task" [ 1103.224874] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.233257] env[62460]: DEBUG oslo_vmware.api [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314136, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.709836] env[62460]: DEBUG oslo_vmware.api [None req-e10c6ab1-73a8-41d7-bbe8-8c1344df9372 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314135, 'name': PowerOnVM_Task, 'duration_secs': 1.730312} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.710177] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-e10c6ab1-73a8-41d7-bbe8-8c1344df9372 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Resumed the VM {{(pid=62460) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 1103.710335] env[62460]: DEBUG nova.compute.manager [None req-e10c6ab1-73a8-41d7-bbe8-8c1344df9372 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1103.711098] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af8d5e33-9001-45cf-8d7a-cb56496b2ab4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.733588] env[62460]: DEBUG oslo_vmware.api [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314136, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.945350] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 61f4666e-0588-4a00-a969-2349148caef2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1103.945645] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1103.946101] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 00597ea8-88ee-448d-b5b5-2b29dc8683e9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1103.946218] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=62460) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1103.946411] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=62460) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1103.993617] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fc1a048-e5a6-4ba7-9e7d-9f42037fedbd {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.000767] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6e9c458-39ca-47ed-8455-782c8cd9c2c7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.030964] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d2fe2ff-b983-49fe-b863-98b920e359ec {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.037830] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94270c36-6413-491d-9603-c42d86e9b8c9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.050240] env[62460]: DEBUG nova.compute.provider_tree [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1104.235327] env[62460]: DEBUG oslo_vmware.api [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314136, 'name': ReconfigVM_Task, 'duration_secs': 0.976112} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.236090] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Reconfigured VM instance instance-00000067 to attach disk [datastore2] 00597ea8-88ee-448d-b5b5-2b29dc8683e9/00597ea8-88ee-448d-b5b5-2b29dc8683e9.vmdk or device None with type streamOptimized {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1104.236700] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-50265d84-92fe-4974-b5df-bd898e40856f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.242567] env[62460]: DEBUG oslo_vmware.api [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 1104.242567] env[62460]: value = "task-1314137" [ 1104.242567] env[62460]: _type = "Task" [ 1104.242567] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.250236] env[62460]: DEBUG oslo_vmware.api [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314137, 'name': Rename_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.553356] env[62460]: DEBUG nova.scheduler.client.report [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1104.751841] env[62460]: DEBUG oslo_vmware.api [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314137, 'name': Rename_Task, 'duration_secs': 0.222575} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.752762] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1104.752762] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-70a2fbf8-6429-4209-9bfb-89384fd9e01e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.759147] env[62460]: DEBUG oslo_vmware.api [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 1104.759147] env[62460]: value = "task-1314138" [ 1104.759147] env[62460]: _type = "Task" [ 1104.759147] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.766836] env[62460]: DEBUG oslo_vmware.api [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314138, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.058177] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62460) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1105.058433] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.138s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1105.269582] env[62460]: DEBUG oslo_vmware.api [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314138, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.770134] env[62460]: DEBUG oslo_vmware.api [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314138, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.272719] env[62460]: DEBUG oslo_vmware.api [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314138, 'name': PowerOnVM_Task, 'duration_secs': 1.448064} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.273036] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1106.378578] env[62460]: DEBUG nova.compute.manager [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1106.379513] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0fcddc7-6b92-4759-9c88-f563e6c2acf8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.897074] env[62460]: DEBUG oslo_concurrency.lockutils [None req-626b8e7d-0fa9-4165-8840-d12a5c46ccba tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lock "00597ea8-88ee-448d-b5b5-2b29dc8683e9" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 22.102s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1106.897886] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "00597ea8-88ee-448d-b5b5-2b29dc8683e9" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 4.550s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1106.898100] env[62460]: INFO nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] During sync_power_state the instance has a pending task (spawning). Skip. [ 1106.898325] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "00597ea8-88ee-448d-b5b5-2b29dc8683e9" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1109.064064] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7b3b47df-e8d4-408d-871d-cca5f1805998 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Acquiring lock "c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1109.064452] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7b3b47df-e8d4-408d-871d-cca5f1805998 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Lock "c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1109.064452] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7b3b47df-e8d4-408d-871d-cca5f1805998 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Acquiring lock "c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1109.064534] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7b3b47df-e8d4-408d-871d-cca5f1805998 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Lock "c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1109.064745] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7b3b47df-e8d4-408d-871d-cca5f1805998 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Lock "c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1109.067040] env[62460]: INFO nova.compute.manager [None req-7b3b47df-e8d4-408d-871d-cca5f1805998 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Terminating instance [ 1109.068616] env[62460]: DEBUG nova.compute.manager [None req-7b3b47df-e8d4-408d-871d-cca5f1805998 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1109.068810] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7b3b47df-e8d4-408d-871d-cca5f1805998 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1109.069653] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ffe5558-2f71-4a77-8551-0915e37b3512 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.077380] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b3b47df-e8d4-408d-871d-cca5f1805998 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1109.077874] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3a015b94-efb0-4c11-b1ce-067ad69c604d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.083652] env[62460]: DEBUG oslo_vmware.api [None req-7b3b47df-e8d4-408d-871d-cca5f1805998 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Waiting for the task: (returnval){ [ 1109.083652] env[62460]: value = "task-1314139" [ 1109.083652] env[62460]: _type = "Task" [ 1109.083652] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.091427] env[62460]: DEBUG oslo_vmware.api [None req-7b3b47df-e8d4-408d-871d-cca5f1805998 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314139, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.594440] env[62460]: DEBUG oslo_vmware.api [None req-7b3b47df-e8d4-408d-871d-cca5f1805998 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314139, 'name': PowerOffVM_Task, 'duration_secs': 0.209065} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.594676] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b3b47df-e8d4-408d-871d-cca5f1805998 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1109.594831] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7b3b47df-e8d4-408d-871d-cca5f1805998 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1109.595098] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-acb6f069-ef0f-4c48-84ec-5fa1f45a9d73 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.658580] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7b3b47df-e8d4-408d-871d-cca5f1805998 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1109.658834] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7b3b47df-e8d4-408d-871d-cca5f1805998 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Deleting contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1109.659011] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b3b47df-e8d4-408d-871d-cca5f1805998 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Deleting the datastore file [datastore1] c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1109.659292] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c7e3fa5c-5f2c-4337-b3de-e27a4488b087 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.665716] env[62460]: DEBUG oslo_vmware.api [None req-7b3b47df-e8d4-408d-871d-cca5f1805998 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Waiting for the task: (returnval){ [ 1109.665716] env[62460]: value = "task-1314141" [ 1109.665716] env[62460]: _type = "Task" [ 1109.665716] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.672914] env[62460]: DEBUG oslo_vmware.api [None req-7b3b47df-e8d4-408d-871d-cca5f1805998 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314141, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.175661] env[62460]: DEBUG oslo_vmware.api [None req-7b3b47df-e8d4-408d-871d-cca5f1805998 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Task: {'id': task-1314141, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.143026} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.176128] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b3b47df-e8d4-408d-871d-cca5f1805998 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1110.176128] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7b3b47df-e8d4-408d-871d-cca5f1805998 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Deleted contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1110.176277] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7b3b47df-e8d4-408d-871d-cca5f1805998 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1110.176488] env[62460]: INFO nova.compute.manager [None req-7b3b47df-e8d4-408d-871d-cca5f1805998 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1110.176780] env[62460]: DEBUG oslo.service.loopingcall [None req-7b3b47df-e8d4-408d-871d-cca5f1805998 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1110.176982] env[62460]: DEBUG nova.compute.manager [-] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1110.177092] env[62460]: DEBUG nova.network.neutron [-] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1110.469267] env[62460]: DEBUG nova.compute.manager [req-18f02302-dde1-4b56-84a3-9b05d621db13 req-5a335a44-16fb-4fb7-a22e-41d2a9301dd1 service nova] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Received event network-vif-deleted-d15647d8-f541-4169-bea9-0937db3301b1 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1110.470015] env[62460]: INFO nova.compute.manager [req-18f02302-dde1-4b56-84a3-9b05d621db13 req-5a335a44-16fb-4fb7-a22e-41d2a9301dd1 service nova] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Neutron deleted interface d15647d8-f541-4169-bea9-0937db3301b1; detaching it from the instance and deleting it from the info cache [ 1110.470015] env[62460]: DEBUG nova.network.neutron [req-18f02302-dde1-4b56-84a3-9b05d621db13 req-5a335a44-16fb-4fb7-a22e-41d2a9301dd1 service nova] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1110.949590] env[62460]: DEBUG nova.network.neutron [-] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1110.972188] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bbb8f3cf-d598-429f-8e80-9b5d3f364fa7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.981182] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f9d92b3-2efb-4541-a754-38c3a96acd9e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.005226] env[62460]: DEBUG nova.compute.manager [req-18f02302-dde1-4b56-84a3-9b05d621db13 req-5a335a44-16fb-4fb7-a22e-41d2a9301dd1 service nova] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Detach interface failed, port_id=d15647d8-f541-4169-bea9-0937db3301b1, reason: Instance c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac could not be found. {{(pid=62460) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1111.452673] env[62460]: INFO nova.compute.manager [-] [instance: c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac] Took 1.28 seconds to deallocate network for instance. [ 1111.959759] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7b3b47df-e8d4-408d-871d-cca5f1805998 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1111.960127] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7b3b47df-e8d4-408d-871d-cca5f1805998 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1111.960415] env[62460]: DEBUG nova.objects.instance [None req-7b3b47df-e8d4-408d-871d-cca5f1805998 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Lazy-loading 'resources' on Instance uuid c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1112.512272] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebe6feb1-dff9-4781-9dae-fcb0a95c6afd {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.519500] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93a017e4-8553-4be0-84d4-e06e5387828d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.548503] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f64178b-92a5-4574-8cf0-a5835f16524c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.555117] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4ac7e30-14ec-4000-bf01-0069894f5fb7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.567472] env[62460]: DEBUG nova.compute.provider_tree [None req-7b3b47df-e8d4-408d-871d-cca5f1805998 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1113.070563] env[62460]: DEBUG nova.scheduler.client.report [None req-7b3b47df-e8d4-408d-871d-cca5f1805998 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1113.577648] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7b3b47df-e8d4-408d-871d-cca5f1805998 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.617s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1113.593436] env[62460]: INFO nova.scheduler.client.report [None req-7b3b47df-e8d4-408d-871d-cca5f1805998 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Deleted allocations for instance c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac [ 1113.918807] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "65024385-6ead-4664-9bf6-1080de47e17d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.919052] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "65024385-6ead-4664-9bf6-1080de47e17d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1114.107326] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7b3b47df-e8d4-408d-871d-cca5f1805998 tempest-ServersNegativeTestJSON-227427144 tempest-ServersNegativeTestJSON-227427144-project-member] Lock "c87fa03b-3f3d-4bff-bfcc-5269fca0b3ac" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.043s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1114.421167] env[62460]: DEBUG nova.compute.manager [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1114.946201] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1114.946480] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1114.948424] env[62460]: INFO nova.compute.claims [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1116.013743] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98cb8f59-1081-460e-a5fb-c1f09ae9bb53 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.025231] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-509ca308-3c54-4270-9765-682e848b2b5e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.056195] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f42d0d84-b369-49f4-91ca-7e87231941c4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.063149] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d9ce341-eb15-40d9-8a11-c62a148a6d94 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.075967] env[62460]: DEBUG nova.compute.provider_tree [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1116.579264] env[62460]: DEBUG nova.scheduler.client.report [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1117.084513] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.138s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1117.085229] env[62460]: DEBUG nova.compute.manager [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1117.590192] env[62460]: DEBUG nova.compute.utils [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1117.591672] env[62460]: DEBUG nova.compute.manager [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1117.591863] env[62460]: DEBUG nova.network.neutron [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1117.653700] env[62460]: DEBUG nova.policy [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a73816ab645e4cd49fabdbc4eeb31f97', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eee8bf23869e4680aec16e37f972175f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 1117.923528] env[62460]: DEBUG nova.network.neutron [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Successfully created port: 79142526-ff67-47b8-9d7d-2e9720cd65ef {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1118.095605] env[62460]: DEBUG nova.compute.manager [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1119.104616] env[62460]: DEBUG nova.compute.manager [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1119.128818] env[62460]: DEBUG nova.virt.hardware [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1119.129080] env[62460]: DEBUG nova.virt.hardware [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1119.129286] env[62460]: DEBUG nova.virt.hardware [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1119.129489] env[62460]: DEBUG nova.virt.hardware [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1119.129648] env[62460]: DEBUG nova.virt.hardware [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1119.129806] env[62460]: DEBUG nova.virt.hardware [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1119.130028] env[62460]: DEBUG nova.virt.hardware [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1119.130206] env[62460]: DEBUG nova.virt.hardware [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1119.130410] env[62460]: DEBUG nova.virt.hardware [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1119.130587] env[62460]: DEBUG nova.virt.hardware [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1119.130793] env[62460]: DEBUG nova.virt.hardware [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1119.131683] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e95c197-ce32-46c4-9512-19f542fa82a9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.139513] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1426ac60-a44d-49cb-bb98-dcc3be038213 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.289095] env[62460]: DEBUG nova.compute.manager [req-bf9e0d69-ef4d-4ece-b068-b775f48db4aa req-dd913cc3-6d5a-41a3-8c62-25d4b9b85348 service nova] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Received event network-vif-plugged-79142526-ff67-47b8-9d7d-2e9720cd65ef {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1119.289331] env[62460]: DEBUG oslo_concurrency.lockutils [req-bf9e0d69-ef4d-4ece-b068-b775f48db4aa req-dd913cc3-6d5a-41a3-8c62-25d4b9b85348 service nova] Acquiring lock "65024385-6ead-4664-9bf6-1080de47e17d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1119.289556] env[62460]: DEBUG oslo_concurrency.lockutils [req-bf9e0d69-ef4d-4ece-b068-b775f48db4aa req-dd913cc3-6d5a-41a3-8c62-25d4b9b85348 service nova] Lock "65024385-6ead-4664-9bf6-1080de47e17d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1119.289733] env[62460]: DEBUG oslo_concurrency.lockutils [req-bf9e0d69-ef4d-4ece-b068-b775f48db4aa req-dd913cc3-6d5a-41a3-8c62-25d4b9b85348 service nova] Lock "65024385-6ead-4664-9bf6-1080de47e17d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1119.289908] env[62460]: DEBUG nova.compute.manager [req-bf9e0d69-ef4d-4ece-b068-b775f48db4aa req-dd913cc3-6d5a-41a3-8c62-25d4b9b85348 service nova] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] No waiting events found dispatching network-vif-plugged-79142526-ff67-47b8-9d7d-2e9720cd65ef {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1119.290094] env[62460]: WARNING nova.compute.manager [req-bf9e0d69-ef4d-4ece-b068-b775f48db4aa req-dd913cc3-6d5a-41a3-8c62-25d4b9b85348 service nova] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Received unexpected event network-vif-plugged-79142526-ff67-47b8-9d7d-2e9720cd65ef for instance with vm_state building and task_state spawning. [ 1119.368624] env[62460]: DEBUG nova.network.neutron [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Successfully updated port: 79142526-ff67-47b8-9d7d-2e9720cd65ef {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1119.874425] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "refresh_cache-65024385-6ead-4664-9bf6-1080de47e17d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1119.874564] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquired lock "refresh_cache-65024385-6ead-4664-9bf6-1080de47e17d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1119.874716] env[62460]: DEBUG nova.network.neutron [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1120.405878] env[62460]: DEBUG nova.network.neutron [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1120.530894] env[62460]: DEBUG nova.network.neutron [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Updating instance_info_cache with network_info: [{"id": "79142526-ff67-47b8-9d7d-2e9720cd65ef", "address": "fa:16:3e:e8:ca:96", "network": {"id": "1dfc278d-e89e-4caf-9cfb-c302106cf974", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-868970047-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eee8bf23869e4680aec16e37f972175f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79142526-ff", "ovs_interfaceid": "79142526-ff67-47b8-9d7d-2e9720cd65ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1121.033505] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Releasing lock "refresh_cache-65024385-6ead-4664-9bf6-1080de47e17d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1121.033848] env[62460]: DEBUG nova.compute.manager [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Instance network_info: |[{"id": "79142526-ff67-47b8-9d7d-2e9720cd65ef", "address": "fa:16:3e:e8:ca:96", "network": {"id": "1dfc278d-e89e-4caf-9cfb-c302106cf974", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-868970047-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eee8bf23869e4680aec16e37f972175f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79142526-ff", "ovs_interfaceid": "79142526-ff67-47b8-9d7d-2e9720cd65ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1121.034325] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e8:ca:96', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '418ddd3d-5f64-407e-8e0c-c8b81639bee9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '79142526-ff67-47b8-9d7d-2e9720cd65ef', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1121.041668] env[62460]: DEBUG oslo.service.loopingcall [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1121.041866] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1121.042112] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9ad92789-d70b-447f-822a-92061b8aa7a6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.062471] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1121.062471] env[62460]: value = "task-1314142" [ 1121.062471] env[62460]: _type = "Task" [ 1121.062471] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.069816] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1314142, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.317191] env[62460]: DEBUG nova.compute.manager [req-d6dff14a-532d-435c-a118-6344a789bdd0 req-18ceee42-434e-441f-a1a7-6629857a42fb service nova] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Received event network-changed-79142526-ff67-47b8-9d7d-2e9720cd65ef {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1121.317361] env[62460]: DEBUG nova.compute.manager [req-d6dff14a-532d-435c-a118-6344a789bdd0 req-18ceee42-434e-441f-a1a7-6629857a42fb service nova] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Refreshing instance network info cache due to event network-changed-79142526-ff67-47b8-9d7d-2e9720cd65ef. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1121.317586] env[62460]: DEBUG oslo_concurrency.lockutils [req-d6dff14a-532d-435c-a118-6344a789bdd0 req-18ceee42-434e-441f-a1a7-6629857a42fb service nova] Acquiring lock "refresh_cache-65024385-6ead-4664-9bf6-1080de47e17d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1121.317735] env[62460]: DEBUG oslo_concurrency.lockutils [req-d6dff14a-532d-435c-a118-6344a789bdd0 req-18ceee42-434e-441f-a1a7-6629857a42fb service nova] Acquired lock "refresh_cache-65024385-6ead-4664-9bf6-1080de47e17d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1121.317907] env[62460]: DEBUG nova.network.neutron [req-d6dff14a-532d-435c-a118-6344a789bdd0 req-18ceee42-434e-441f-a1a7-6629857a42fb service nova] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Refreshing network info cache for port 79142526-ff67-47b8-9d7d-2e9720cd65ef {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1121.572487] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1314142, 'name': CreateVM_Task, 'duration_secs': 0.302856} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.572868] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1121.573322] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1121.573491] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1121.573819] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1121.574085] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c047823-aa69-4456-a5ca-cd00575740ca {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.578800] env[62460]: DEBUG oslo_vmware.api [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1121.578800] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]5235304c-2f9b-db44-a3ab-e2ca8092cc58" [ 1121.578800] env[62460]: _type = "Task" [ 1121.578800] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.586062] env[62460]: DEBUG oslo_vmware.api [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5235304c-2f9b-db44-a3ab-e2ca8092cc58, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.999357] env[62460]: DEBUG nova.network.neutron [req-d6dff14a-532d-435c-a118-6344a789bdd0 req-18ceee42-434e-441f-a1a7-6629857a42fb service nova] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Updated VIF entry in instance network info cache for port 79142526-ff67-47b8-9d7d-2e9720cd65ef. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1121.999717] env[62460]: DEBUG nova.network.neutron [req-d6dff14a-532d-435c-a118-6344a789bdd0 req-18ceee42-434e-441f-a1a7-6629857a42fb service nova] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Updating instance_info_cache with network_info: [{"id": "79142526-ff67-47b8-9d7d-2e9720cd65ef", "address": "fa:16:3e:e8:ca:96", "network": {"id": "1dfc278d-e89e-4caf-9cfb-c302106cf974", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-868970047-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eee8bf23869e4680aec16e37f972175f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79142526-ff", "ovs_interfaceid": "79142526-ff67-47b8-9d7d-2e9720cd65ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1122.088482] env[62460]: DEBUG oslo_vmware.api [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5235304c-2f9b-db44-a3ab-e2ca8092cc58, 'name': SearchDatastore_Task, 'duration_secs': 0.01468} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.088786] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1122.089039] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1122.089300] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1122.089438] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1122.089625] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1122.089876] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-88c78b38-77d4-4398-9ba5-dd1ff97718a0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.097666] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1122.097817] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1122.098498] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0225d907-b0a4-4551-85d7-555fdf2fc7c0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.103045] env[62460]: DEBUG oslo_vmware.api [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1122.103045] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52f27e33-00f8-7c67-e46b-67b563bc6fdd" [ 1122.103045] env[62460]: _type = "Task" [ 1122.103045] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.109960] env[62460]: DEBUG oslo_vmware.api [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52f27e33-00f8-7c67-e46b-67b563bc6fdd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.502893] env[62460]: DEBUG oslo_concurrency.lockutils [req-d6dff14a-532d-435c-a118-6344a789bdd0 req-18ceee42-434e-441f-a1a7-6629857a42fb service nova] Releasing lock "refresh_cache-65024385-6ead-4664-9bf6-1080de47e17d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1122.613496] env[62460]: DEBUG oslo_vmware.api [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52f27e33-00f8-7c67-e46b-67b563bc6fdd, 'name': SearchDatastore_Task, 'duration_secs': 0.007619} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.614293] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-26d74973-dfdc-4ec5-ae40-855af7d80127 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.619178] env[62460]: DEBUG oslo_vmware.api [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1122.619178] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]5264d97c-1c7a-ff64-7947-ec27a5fbd030" [ 1122.619178] env[62460]: _type = "Task" [ 1122.619178] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.627788] env[62460]: DEBUG oslo_vmware.api [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5264d97c-1c7a-ff64-7947-ec27a5fbd030, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.129602] env[62460]: DEBUG oslo_vmware.api [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5264d97c-1c7a-ff64-7947-ec27a5fbd030, 'name': SearchDatastore_Task, 'duration_secs': 0.009122} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.129880] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1123.130154] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 65024385-6ead-4664-9bf6-1080de47e17d/65024385-6ead-4664-9bf6-1080de47e17d.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1123.130518] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-29630d07-7920-4e6e-92d4-5429d1de81a3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.137560] env[62460]: DEBUG oslo_vmware.api [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1123.137560] env[62460]: value = "task-1314143" [ 1123.137560] env[62460]: _type = "Task" [ 1123.137560] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.144902] env[62460]: DEBUG oslo_vmware.api [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314143, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.647568] env[62460]: DEBUG oslo_vmware.api [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314143, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.476491} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.647944] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 65024385-6ead-4664-9bf6-1080de47e17d/65024385-6ead-4664-9bf6-1080de47e17d.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1123.648079] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1123.648340] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-88ebbb39-03ad-47b8-bc8b-f3a3daffc615 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.654065] env[62460]: DEBUG oslo_vmware.api [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1123.654065] env[62460]: value = "task-1314144" [ 1123.654065] env[62460]: _type = "Task" [ 1123.654065] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.660995] env[62460]: DEBUG oslo_vmware.api [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314144, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.164264] env[62460]: DEBUG oslo_vmware.api [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314144, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061991} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.164494] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1124.165245] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c3c677f-7872-458b-9276-033624cc31f8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.187289] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] 65024385-6ead-4664-9bf6-1080de47e17d/65024385-6ead-4664-9bf6-1080de47e17d.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1124.187564] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e1c5f96b-0011-411d-877f-488340db3f3d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.206717] env[62460]: DEBUG oslo_vmware.api [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1124.206717] env[62460]: value = "task-1314145" [ 1124.206717] env[62460]: _type = "Task" [ 1124.206717] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.214311] env[62460]: DEBUG oslo_vmware.api [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314145, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.716716] env[62460]: DEBUG oslo_vmware.api [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314145, 'name': ReconfigVM_Task, 'duration_secs': 0.278874} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.717094] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Reconfigured VM instance instance-0000006a to attach disk [datastore1] 65024385-6ead-4664-9bf6-1080de47e17d/65024385-6ead-4664-9bf6-1080de47e17d.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1124.717664] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-408f5485-8f85-458a-8034-160faf3cb0b7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.725056] env[62460]: DEBUG oslo_vmware.api [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1124.725056] env[62460]: value = "task-1314146" [ 1124.725056] env[62460]: _type = "Task" [ 1124.725056] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.731474] env[62460]: DEBUG oslo_vmware.api [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314146, 'name': Rename_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.234287] env[62460]: DEBUG oslo_vmware.api [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314146, 'name': Rename_Task, 'duration_secs': 0.187791} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.234582] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1125.234832] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-368a5b2d-86f1-462a-afbe-2b949e444e64 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.241343] env[62460]: DEBUG oslo_vmware.api [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1125.241343] env[62460]: value = "task-1314147" [ 1125.241343] env[62460]: _type = "Task" [ 1125.241343] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.248330] env[62460]: DEBUG oslo_vmware.api [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314147, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.751900] env[62460]: DEBUG oslo_vmware.api [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314147, 'name': PowerOnVM_Task, 'duration_secs': 0.444215} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.752283] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1125.752392] env[62460]: INFO nova.compute.manager [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Took 6.65 seconds to spawn the instance on the hypervisor. [ 1125.752585] env[62460]: DEBUG nova.compute.manager [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1125.753343] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b314f9b-b92d-4acb-a447-c3bf0e2bc89f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.271895] env[62460]: INFO nova.compute.manager [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Took 11.34 seconds to build instance. [ 1126.773809] env[62460]: DEBUG oslo_concurrency.lockutils [None req-0ce829cc-dd8f-49ed-b211-039ad289722a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "65024385-6ead-4664-9bf6-1080de47e17d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.855s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1127.254321] env[62460]: INFO nova.compute.manager [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Rebuilding instance [ 1127.366793] env[62460]: DEBUG nova.compute.manager [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1127.367724] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45e7de2d-295b-4714-ad3e-03d5d8564470 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.878314] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1127.878676] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7a88c153-3b87-4893-95e9-d6f8ee3f0004 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.885989] env[62460]: DEBUG oslo_vmware.api [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1127.885989] env[62460]: value = "task-1314148" [ 1127.885989] env[62460]: _type = "Task" [ 1127.885989] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.893986] env[62460]: DEBUG oslo_vmware.api [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314148, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.395807] env[62460]: DEBUG oslo_vmware.api [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314148, 'name': PowerOffVM_Task, 'duration_secs': 0.209757} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.396085] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1128.396314] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1128.397055] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77143fa1-69d2-411f-915e-e21ef725571d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.403278] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1128.403496] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dbcef558-ccc4-4a70-adbd-51f325e822b0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.957038] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1129.957419] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Deleting contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1129.957598] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Deleting the datastore file [datastore1] 65024385-6ead-4664-9bf6-1080de47e17d {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1129.957887] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-28641e40-6dc2-42de-8202-7e54d8110026 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.964062] env[62460]: DEBUG oslo_vmware.api [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1129.964062] env[62460]: value = "task-1314150" [ 1129.964062] env[62460]: _type = "Task" [ 1129.964062] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.971733] env[62460]: DEBUG oslo_vmware.api [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314150, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.474406] env[62460]: DEBUG oslo_vmware.api [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314150, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.136832} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.474711] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1130.474911] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Deleted contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1130.475112] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1131.504280] env[62460]: DEBUG nova.virt.hardware [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1131.504557] env[62460]: DEBUG nova.virt.hardware [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1131.504700] env[62460]: DEBUG nova.virt.hardware [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1131.504888] env[62460]: DEBUG nova.virt.hardware [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1131.505065] env[62460]: DEBUG nova.virt.hardware [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1131.505230] env[62460]: DEBUG nova.virt.hardware [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1131.505445] env[62460]: DEBUG nova.virt.hardware [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1131.505615] env[62460]: DEBUG nova.virt.hardware [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1131.505790] env[62460]: DEBUG nova.virt.hardware [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1131.505957] env[62460]: DEBUG nova.virt.hardware [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1131.506154] env[62460]: DEBUG nova.virt.hardware [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1131.507021] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cace6dc-cae5-4462-8347-e6ae577cebb7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.514676] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1da63f0f-a8ee-4e8e-90ee-d3b5d7e5b061 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.528919] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e8:ca:96', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '418ddd3d-5f64-407e-8e0c-c8b81639bee9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '79142526-ff67-47b8-9d7d-2e9720cd65ef', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1131.536139] env[62460]: DEBUG oslo.service.loopingcall [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1131.536387] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1131.536594] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d6e24b80-bed6-4817-a57d-caf77a1c5e48 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.554654] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1131.554654] env[62460]: value = "task-1314151" [ 1131.554654] env[62460]: _type = "Task" [ 1131.554654] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.561744] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1314151, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.063964] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1314151, 'name': CreateVM_Task, 'duration_secs': 0.279252} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.064215] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1132.064793] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1132.064970] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1132.065322] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1132.065597] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f4e6cf59-14b3-4e30-8bfb-7cb670861644 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.069614] env[62460]: DEBUG oslo_vmware.api [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1132.069614] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]528a2630-4333-cea2-343a-e3cefb59e971" [ 1132.069614] env[62460]: _type = "Task" [ 1132.069614] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.076804] env[62460]: DEBUG oslo_vmware.api [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]528a2630-4333-cea2-343a-e3cefb59e971, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.580144] env[62460]: DEBUG oslo_vmware.api [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]528a2630-4333-cea2-343a-e3cefb59e971, 'name': SearchDatastore_Task, 'duration_secs': 0.009187} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.580566] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1132.580723] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1132.580963] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1132.581133] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1132.581323] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1132.581587] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-00913c7e-2178-4af2-850d-5a100268b638 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.590324] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1132.590585] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1132.591259] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-00afea7c-fde7-4b41-823c-252af577fb4a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.596138] env[62460]: DEBUG oslo_vmware.api [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1132.596138] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]521d3ebf-fad7-5998-db95-acbc6d70a8c5" [ 1132.596138] env[62460]: _type = "Task" [ 1132.596138] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.603275] env[62460]: DEBUG oslo_vmware.api [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]521d3ebf-fad7-5998-db95-acbc6d70a8c5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.106522] env[62460]: DEBUG oslo_vmware.api [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]521d3ebf-fad7-5998-db95-acbc6d70a8c5, 'name': SearchDatastore_Task, 'duration_secs': 0.008488} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.107338] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-70756a48-8834-4f72-b9d7-34f9d680b922 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.112285] env[62460]: DEBUG oslo_vmware.api [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1133.112285] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52a0403a-f134-d5cd-379d-792cc522f66d" [ 1133.112285] env[62460]: _type = "Task" [ 1133.112285] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.119762] env[62460]: DEBUG oslo_vmware.api [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52a0403a-f134-d5cd-379d-792cc522f66d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.621837] env[62460]: DEBUG oslo_vmware.api [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52a0403a-f134-d5cd-379d-792cc522f66d, 'name': SearchDatastore_Task, 'duration_secs': 0.00897} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.622251] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1133.622337] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 65024385-6ead-4664-9bf6-1080de47e17d/65024385-6ead-4664-9bf6-1080de47e17d.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1133.622654] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b1d5d8d4-33d1-49fe-9e30-4ab2e1dc385e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.629310] env[62460]: DEBUG oslo_vmware.api [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1133.629310] env[62460]: value = "task-1314152" [ 1133.629310] env[62460]: _type = "Task" [ 1133.629310] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.636995] env[62460]: DEBUG oslo_vmware.api [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314152, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.139206] env[62460]: DEBUG oslo_vmware.api [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314152, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.467883} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.139514] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 65024385-6ead-4664-9bf6-1080de47e17d/65024385-6ead-4664-9bf6-1080de47e17d.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1134.139760] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1134.140034] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-45962804-a420-47a3-86a7-e9c1fe06e94f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.145841] env[62460]: DEBUG oslo_vmware.api [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1134.145841] env[62460]: value = "task-1314153" [ 1134.145841] env[62460]: _type = "Task" [ 1134.145841] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.153363] env[62460]: DEBUG oslo_vmware.api [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314153, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.655725] env[62460]: DEBUG oslo_vmware.api [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314153, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.411148} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.656130] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1134.656771] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbee7deb-6754-41bd-a549-7c03d6ffb804 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.678731] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] 65024385-6ead-4664-9bf6-1080de47e17d/65024385-6ead-4664-9bf6-1080de47e17d.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1134.679037] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-85d223d5-4139-4223-81e9-08d6e3472888 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.697840] env[62460]: DEBUG oslo_vmware.api [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1134.697840] env[62460]: value = "task-1314154" [ 1134.697840] env[62460]: _type = "Task" [ 1134.697840] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.707026] env[62460]: DEBUG oslo_vmware.api [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314154, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.208369] env[62460]: DEBUG oslo_vmware.api [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314154, 'name': ReconfigVM_Task, 'duration_secs': 0.266911} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.208641] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Reconfigured VM instance instance-0000006a to attach disk [datastore1] 65024385-6ead-4664-9bf6-1080de47e17d/65024385-6ead-4664-9bf6-1080de47e17d.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1135.209262] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-52308482-84a6-4c2d-bb53-29ff0b928d12 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.215424] env[62460]: DEBUG oslo_vmware.api [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1135.215424] env[62460]: value = "task-1314155" [ 1135.215424] env[62460]: _type = "Task" [ 1135.215424] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.222334] env[62460]: DEBUG oslo_vmware.api [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314155, 'name': Rename_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.724925] env[62460]: DEBUG oslo_vmware.api [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314155, 'name': Rename_Task, 'duration_secs': 0.132997} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.725266] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1135.725496] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-383033a0-e7c3-422d-b710-ceb78401311d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.731978] env[62460]: DEBUG oslo_vmware.api [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1135.731978] env[62460]: value = "task-1314156" [ 1135.731978] env[62460]: _type = "Task" [ 1135.731978] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.739082] env[62460]: DEBUG oslo_vmware.api [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314156, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.097212] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1136.097470] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1136.097584] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Starting heal instance info cache {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1136.241681] env[62460]: DEBUG oslo_vmware.api [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314156, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.652523] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Acquiring lock "refresh_cache-00597ea8-88ee-448d-b5b5-2b29dc8683e9" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1136.652717] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Acquired lock "refresh_cache-00597ea8-88ee-448d-b5b5-2b29dc8683e9" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1136.652876] env[62460]: DEBUG nova.network.neutron [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Forcefully refreshing network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1136.742036] env[62460]: DEBUG oslo_vmware.api [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314156, 'name': PowerOnVM_Task, 'duration_secs': 0.613519} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.742352] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1136.742539] env[62460]: DEBUG nova.compute.manager [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1136.743344] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c336d44-7715-4e22-94fe-befcba9b8f27 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.260577] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1137.260577] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1137.260577] env[62460]: DEBUG nova.objects.instance [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62460) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1137.871916] env[62460]: DEBUG nova.network.neutron [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Updating instance_info_cache with network_info: [{"id": "c591ec72-086f-44a0-8cf2-3e6719700505", "address": "fa:16:3e:30:d9:88", "network": {"id": "b5cb4ee0-4562-45be-a25d-51c3a5932efd", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-823340443-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.156", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3d2d1c48ec14121a2e8c9b3f800a949", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55c757ac-f8b2-466d-b634-07dbd100b312", "external-id": "nsx-vlan-transportzone-159", "segmentation_id": 159, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc591ec72-08", "ovs_interfaceid": "c591ec72-086f-44a0-8cf2-3e6719700505", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1138.269710] env[62460]: DEBUG oslo_concurrency.lockutils [None req-3120fbdb-193b-432e-a028-7758c9fe8818 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.009s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1138.374463] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Releasing lock "refresh_cache-00597ea8-88ee-448d-b5b5-2b29dc8683e9" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1138.374463] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Updated the network info_cache for instance {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1138.374463] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1138.374647] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1138.374747] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1138.374899] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1138.375055] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1138.375208] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1138.375340] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62460) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1138.741461] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7fbe095f-e66f-431d-8915-e112fe70c36c tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "65024385-6ead-4664-9bf6-1080de47e17d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1138.741758] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7fbe095f-e66f-431d-8915-e112fe70c36c tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "65024385-6ead-4664-9bf6-1080de47e17d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1138.741985] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7fbe095f-e66f-431d-8915-e112fe70c36c tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "65024385-6ead-4664-9bf6-1080de47e17d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1138.742200] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7fbe095f-e66f-431d-8915-e112fe70c36c tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "65024385-6ead-4664-9bf6-1080de47e17d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1138.742378] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7fbe095f-e66f-431d-8915-e112fe70c36c tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "65024385-6ead-4664-9bf6-1080de47e17d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1138.744632] env[62460]: INFO nova.compute.manager [None req-7fbe095f-e66f-431d-8915-e112fe70c36c tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Terminating instance [ 1138.746367] env[62460]: DEBUG nova.compute.manager [None req-7fbe095f-e66f-431d-8915-e112fe70c36c tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1138.746566] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7fbe095f-e66f-431d-8915-e112fe70c36c tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1138.747419] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efc5e6b7-7549-41fe-ae0b-7057e773d096 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.754466] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fbe095f-e66f-431d-8915-e112fe70c36c tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1138.754695] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8e7ad191-e9a7-4b8b-9c9a-a2f6b02c0152 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.760832] env[62460]: DEBUG oslo_vmware.api [None req-7fbe095f-e66f-431d-8915-e112fe70c36c tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1138.760832] env[62460]: value = "task-1314157" [ 1138.760832] env[62460]: _type = "Task" [ 1138.760832] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.768110] env[62460]: DEBUG oslo_vmware.api [None req-7fbe095f-e66f-431d-8915-e112fe70c36c tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314157, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.271239] env[62460]: DEBUG oslo_vmware.api [None req-7fbe095f-e66f-431d-8915-e112fe70c36c tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314157, 'name': PowerOffVM_Task, 'duration_secs': 0.181518} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.271619] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fbe095f-e66f-431d-8915-e112fe70c36c tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1139.271659] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7fbe095f-e66f-431d-8915-e112fe70c36c tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1139.271907] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3be01fa0-34db-4e2e-b85d-300b8d9193c6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.868435] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1140.371968] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1140.372266] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1140.372394] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1140.372551] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62460) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1140.373481] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2598e706-488c-4143-b433-a39c248a048d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.381811] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe045d0b-e9bb-4b5b-9347-73a241a705a0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.395165] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17829197-43da-4283-8c60-a2eeae08e917 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.401169] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9874e01b-f359-40d9-bfec-d8cfcbef7729 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.431058] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180916MB free_disk=135GB free_vcpus=48 pci_devices=None {{(pid=62460) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1140.431226] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1140.431403] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1141.476868] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 61f4666e-0588-4a00-a969-2349148caef2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1141.477149] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 00597ea8-88ee-448d-b5b5-2b29dc8683e9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1141.477198] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 65024385-6ead-4664-9bf6-1080de47e17d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1141.477353] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=62460) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1141.477530] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=62460) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1141.525260] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68a751b8-5387-4fd0-86e6-456359c91178 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.532330] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98895ce0-70a6-4b54-9a5f-d31f9b46c950 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.563869] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb6bce72-8114-4f27-b423-bab125b0bf33 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.570410] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b49d99e-881e-470c-94ca-bced8c7a1e14 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.582774] env[62460]: DEBUG nova.compute.provider_tree [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1141.882766] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7fbe095f-e66f-431d-8915-e112fe70c36c tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1141.883040] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7fbe095f-e66f-431d-8915-e112fe70c36c tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Deleting contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1141.883208] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-7fbe095f-e66f-431d-8915-e112fe70c36c tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Deleting the datastore file [datastore1] 65024385-6ead-4664-9bf6-1080de47e17d {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1141.883498] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-aaf287ff-d611-4710-8033-618f076451a7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.889864] env[62460]: DEBUG oslo_vmware.api [None req-7fbe095f-e66f-431d-8915-e112fe70c36c tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1141.889864] env[62460]: value = "task-1314159" [ 1141.889864] env[62460]: _type = "Task" [ 1141.889864] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.897106] env[62460]: DEBUG oslo_vmware.api [None req-7fbe095f-e66f-431d-8915-e112fe70c36c tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314159, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.086487] env[62460]: DEBUG nova.scheduler.client.report [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1142.398841] env[62460]: DEBUG oslo_vmware.api [None req-7fbe095f-e66f-431d-8915-e112fe70c36c tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314159, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.14422} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.400099] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-7fbe095f-e66f-431d-8915-e112fe70c36c tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1142.400099] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7fbe095f-e66f-431d-8915-e112fe70c36c tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Deleted contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1142.400099] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-7fbe095f-e66f-431d-8915-e112fe70c36c tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1142.400099] env[62460]: INFO nova.compute.manager [None req-7fbe095f-e66f-431d-8915-e112fe70c36c tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Took 3.65 seconds to destroy the instance on the hypervisor. [ 1142.400321] env[62460]: DEBUG oslo.service.loopingcall [None req-7fbe095f-e66f-431d-8915-e112fe70c36c tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1142.400476] env[62460]: DEBUG nova.compute.manager [-] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1142.400539] env[62460]: DEBUG nova.network.neutron [-] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1142.591544] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62460) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1142.591795] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.160s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1142.784654] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a57b29a1-c9bd-43d0-a6f2-6952fdaef8b7 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquiring lock "00597ea8-88ee-448d-b5b5-2b29dc8683e9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1142.784939] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a57b29a1-c9bd-43d0-a6f2-6952fdaef8b7 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lock "00597ea8-88ee-448d-b5b5-2b29dc8683e9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1142.785172] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a57b29a1-c9bd-43d0-a6f2-6952fdaef8b7 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquiring lock "00597ea8-88ee-448d-b5b5-2b29dc8683e9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1142.785367] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a57b29a1-c9bd-43d0-a6f2-6952fdaef8b7 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lock "00597ea8-88ee-448d-b5b5-2b29dc8683e9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1142.785544] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a57b29a1-c9bd-43d0-a6f2-6952fdaef8b7 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lock "00597ea8-88ee-448d-b5b5-2b29dc8683e9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1142.787692] env[62460]: INFO nova.compute.manager [None req-a57b29a1-c9bd-43d0-a6f2-6952fdaef8b7 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Terminating instance [ 1142.789324] env[62460]: DEBUG nova.compute.manager [None req-a57b29a1-c9bd-43d0-a6f2-6952fdaef8b7 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1142.789526] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a57b29a1-c9bd-43d0-a6f2-6952fdaef8b7 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1142.790446] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5589e890-71ce-4e18-9782-d257469739a1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.795195] env[62460]: DEBUG nova.compute.manager [req-349d080e-8af9-4802-8792-47f047336699 req-e2e85d7f-5350-4c70-b9ac-22be7d666b30 service nova] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Received event network-vif-deleted-79142526-ff67-47b8-9d7d-2e9720cd65ef {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1142.795386] env[62460]: INFO nova.compute.manager [req-349d080e-8af9-4802-8792-47f047336699 req-e2e85d7f-5350-4c70-b9ac-22be7d666b30 service nova] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Neutron deleted interface 79142526-ff67-47b8-9d7d-2e9720cd65ef; detaching it from the instance and deleting it from the info cache [ 1142.795563] env[62460]: DEBUG nova.network.neutron [req-349d080e-8af9-4802-8792-47f047336699 req-e2e85d7f-5350-4c70-b9ac-22be7d666b30 service nova] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1142.800854] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-a57b29a1-c9bd-43d0-a6f2-6952fdaef8b7 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1142.801093] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c0b98adb-8ba5-4e32-bbdc-ce375bb5865b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.807033] env[62460]: DEBUG oslo_vmware.api [None req-a57b29a1-c9bd-43d0-a6f2-6952fdaef8b7 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 1142.807033] env[62460]: value = "task-1314160" [ 1142.807033] env[62460]: _type = "Task" [ 1142.807033] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1142.815489] env[62460]: DEBUG oslo_vmware.api [None req-a57b29a1-c9bd-43d0-a6f2-6952fdaef8b7 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314160, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.207841] env[62460]: DEBUG nova.network.neutron [-] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1143.298722] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-56200970-93e0-4f09-9cee-e0f416158aff {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.307897] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-584c7503-5e0a-46be-a82e-012daa2fba04 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.325841] env[62460]: DEBUG oslo_vmware.api [None req-a57b29a1-c9bd-43d0-a6f2-6952fdaef8b7 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314160, 'name': PowerOffVM_Task, 'duration_secs': 0.167108} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.326131] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-a57b29a1-c9bd-43d0-a6f2-6952fdaef8b7 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1143.326310] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a57b29a1-c9bd-43d0-a6f2-6952fdaef8b7 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1143.326583] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-95ca94e6-8271-4461-9296-b28432584fbb {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.332878] env[62460]: DEBUG nova.compute.manager [req-349d080e-8af9-4802-8792-47f047336699 req-e2e85d7f-5350-4c70-b9ac-22be7d666b30 service nova] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Detach interface failed, port_id=79142526-ff67-47b8-9d7d-2e9720cd65ef, reason: Instance 65024385-6ead-4664-9bf6-1080de47e17d could not be found. {{(pid=62460) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1143.385145] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a57b29a1-c9bd-43d0-a6f2-6952fdaef8b7 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1143.385348] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a57b29a1-c9bd-43d0-a6f2-6952fdaef8b7 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Deleting contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1143.385528] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-a57b29a1-c9bd-43d0-a6f2-6952fdaef8b7 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Deleting the datastore file [datastore2] 00597ea8-88ee-448d-b5b5-2b29dc8683e9 {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1143.385789] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7f8d2795-6abe-46f5-aa57-f1f602c50ac9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.391906] env[62460]: DEBUG oslo_vmware.api [None req-a57b29a1-c9bd-43d0-a6f2-6952fdaef8b7 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for the task: (returnval){ [ 1143.391906] env[62460]: value = "task-1314162" [ 1143.391906] env[62460]: _type = "Task" [ 1143.391906] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.399685] env[62460]: DEBUG oslo_vmware.api [None req-a57b29a1-c9bd-43d0-a6f2-6952fdaef8b7 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314162, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.710423] env[62460]: INFO nova.compute.manager [-] [instance: 65024385-6ead-4664-9bf6-1080de47e17d] Took 1.31 seconds to deallocate network for instance. [ 1143.902430] env[62460]: DEBUG oslo_vmware.api [None req-a57b29a1-c9bd-43d0-a6f2-6952fdaef8b7 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Task: {'id': task-1314162, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.166176} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.902707] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-a57b29a1-c9bd-43d0-a6f2-6952fdaef8b7 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1143.902899] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a57b29a1-c9bd-43d0-a6f2-6952fdaef8b7 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Deleted contents of the VM from datastore datastore2 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1143.903095] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-a57b29a1-c9bd-43d0-a6f2-6952fdaef8b7 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1143.903284] env[62460]: INFO nova.compute.manager [None req-a57b29a1-c9bd-43d0-a6f2-6952fdaef8b7 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1143.903530] env[62460]: DEBUG oslo.service.loopingcall [None req-a57b29a1-c9bd-43d0-a6f2-6952fdaef8b7 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1143.903731] env[62460]: DEBUG nova.compute.manager [-] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1143.903827] env[62460]: DEBUG nova.network.neutron [-] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1144.216869] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7fbe095f-e66f-431d-8915-e112fe70c36c tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1144.217189] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7fbe095f-e66f-431d-8915-e112fe70c36c tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1144.217423] env[62460]: DEBUG nova.objects.instance [None req-7fbe095f-e66f-431d-8915-e112fe70c36c tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lazy-loading 'resources' on Instance uuid 65024385-6ead-4664-9bf6-1080de47e17d {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1144.795479] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8d3b595-a3b0-4567-9d59-9860ea24aa6e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.803166] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c439ccd5-e5b5-459d-b964-28662f3f7a9f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.832127] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16e9feca-16d1-461e-9139-2bf7bbb58618 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.839239] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-089c4e65-5a57-44d2-bb26-e0f8a8a88112 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.852195] env[62460]: DEBUG nova.compute.provider_tree [None req-7fbe095f-e66f-431d-8915-e112fe70c36c tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1144.855276] env[62460]: DEBUG nova.compute.manager [req-0f1ba44e-f6b2-4fec-8abe-2c657d9208ad req-2e8fd8fb-bd20-4882-8e4f-91a2019791a0 service nova] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Received event network-vif-deleted-c591ec72-086f-44a0-8cf2-3e6719700505 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1144.855470] env[62460]: INFO nova.compute.manager [req-0f1ba44e-f6b2-4fec-8abe-2c657d9208ad req-2e8fd8fb-bd20-4882-8e4f-91a2019791a0 service nova] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Neutron deleted interface c591ec72-086f-44a0-8cf2-3e6719700505; detaching it from the instance and deleting it from the info cache [ 1144.855647] env[62460]: DEBUG nova.network.neutron [req-0f1ba44e-f6b2-4fec-8abe-2c657d9208ad req-2e8fd8fb-bd20-4882-8e4f-91a2019791a0 service nova] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1144.890387] env[62460]: DEBUG nova.network.neutron [-] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1145.359054] env[62460]: DEBUG nova.scheduler.client.report [None req-7fbe095f-e66f-431d-8915-e112fe70c36c tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1145.361890] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d9019ab3-33b8-4191-9099-21769fe9bccc {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.372869] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2e45a6c-19be-45c9-9ce8-5a571bf44cad {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.395265] env[62460]: INFO nova.compute.manager [-] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Took 1.49 seconds to deallocate network for instance. [ 1145.395610] env[62460]: DEBUG nova.compute.manager [req-0f1ba44e-f6b2-4fec-8abe-2c657d9208ad req-2e8fd8fb-bd20-4882-8e4f-91a2019791a0 service nova] [instance: 00597ea8-88ee-448d-b5b5-2b29dc8683e9] Detach interface failed, port_id=c591ec72-086f-44a0-8cf2-3e6719700505, reason: Instance 00597ea8-88ee-448d-b5b5-2b29dc8683e9 could not be found. {{(pid=62460) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1145.863015] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7fbe095f-e66f-431d-8915-e112fe70c36c tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.646s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1145.901018] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a57b29a1-c9bd-43d0-a6f2-6952fdaef8b7 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1145.901313] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a57b29a1-c9bd-43d0-a6f2-6952fdaef8b7 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1145.901586] env[62460]: DEBUG nova.objects.instance [None req-a57b29a1-c9bd-43d0-a6f2-6952fdaef8b7 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lazy-loading 'resources' on Instance uuid 00597ea8-88ee-448d-b5b5-2b29dc8683e9 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1145.909520] env[62460]: INFO nova.scheduler.client.report [None req-7fbe095f-e66f-431d-8915-e112fe70c36c tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Deleted allocations for instance 65024385-6ead-4664-9bf6-1080de47e17d [ 1146.417038] env[62460]: DEBUG oslo_concurrency.lockutils [None req-7fbe095f-e66f-431d-8915-e112fe70c36c tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "65024385-6ead-4664-9bf6-1080de47e17d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.675s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1146.441825] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f5e1358-84ed-4de9-ab3b-055016872f1b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.449013] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82b88061-95e6-4686-a421-df3b100aef0e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.477490] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d2365da-3718-4f8d-a87c-2c2642339a4d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.484169] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ad59647-4f2e-4249-bcbf-1afe535fbf8f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.497737] env[62460]: DEBUG nova.compute.provider_tree [None req-a57b29a1-c9bd-43d0-a6f2-6952fdaef8b7 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1147.001161] env[62460]: DEBUG nova.scheduler.client.report [None req-a57b29a1-c9bd-43d0-a6f2-6952fdaef8b7 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1147.505896] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a57b29a1-c9bd-43d0-a6f2-6952fdaef8b7 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.604s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1147.526117] env[62460]: INFO nova.scheduler.client.report [None req-a57b29a1-c9bd-43d0-a6f2-6952fdaef8b7 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Deleted allocations for instance 00597ea8-88ee-448d-b5b5-2b29dc8683e9 [ 1148.034665] env[62460]: DEBUG oslo_concurrency.lockutils [None req-a57b29a1-c9bd-43d0-a6f2-6952fdaef8b7 tempest-AttachVolumeShelveTestJSON-60162227 tempest-AttachVolumeShelveTestJSON-60162227-project-member] Lock "00597ea8-88ee-448d-b5b5-2b29dc8683e9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.250s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1148.614226] env[62460]: DEBUG nova.compute.manager [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Stashing vm_state: active {{(pid=62460) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1149.135640] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1149.135907] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1149.641969] env[62460]: INFO nova.compute.claims [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1150.148116] env[62460]: INFO nova.compute.resource_tracker [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Updating resource usage from migration e7a3c104-1792-4b07-9214-3984fb4e1c59 [ 1150.184354] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fe4d1ec-1772-4ba6-b862-74f32a360f7b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.191652] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-840047a0-b53a-41b7-926f-1e951ce95a39 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.221161] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c2a79b3-115a-4472-8999-55635e3c817d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.227827] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eefd4664-3537-44b2-8dee-89cbca8e82f2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.240615] env[62460]: DEBUG nova.compute.provider_tree [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1150.743705] env[62460]: DEBUG nova.scheduler.client.report [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1151.248802] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.113s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1151.249175] env[62460]: INFO nova.compute.manager [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Migrating [ 1151.764138] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "refresh_cache-61f4666e-0588-4a00-a969-2349148caef2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1151.764376] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquired lock "refresh_cache-61f4666e-0588-4a00-a969-2349148caef2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1151.764539] env[62460]: DEBUG nova.network.neutron [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1152.491670] env[62460]: DEBUG nova.network.neutron [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Updating instance_info_cache with network_info: [{"id": "345c87ec-2bb1-420f-bc8c-845a795baad5", "address": "fa:16:3e:43:36:b0", "network": {"id": "1dfc278d-e89e-4caf-9cfb-c302106cf974", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-868970047-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eee8bf23869e4680aec16e37f972175f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap345c87ec-2b", "ovs_interfaceid": "345c87ec-2bb1-420f-bc8c-845a795baad5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1152.995686] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Releasing lock "refresh_cache-61f4666e-0588-4a00-a969-2349148caef2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1154.513028] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19fba307-234f-42ea-b9b4-6a29e770ac4a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.532106] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Updating instance '61f4666e-0588-4a00-a969-2349148caef2' progress to 0 {{(pid=62460) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1155.037957] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1155.038295] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-14da965c-761c-4e10-b705-d761ec25360d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.045653] env[62460]: DEBUG oslo_vmware.api [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1155.045653] env[62460]: value = "task-1314164" [ 1155.045653] env[62460]: _type = "Task" [ 1155.045653] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.053505] env[62460]: DEBUG oslo_vmware.api [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314164, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.555575] env[62460]: DEBUG oslo_vmware.api [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314164, 'name': PowerOffVM_Task, 'duration_secs': 0.171888} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.555965] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1155.556050] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Updating instance '61f4666e-0588-4a00-a969-2349148caef2' progress to 17 {{(pid=62460) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1156.062642] env[62460]: DEBUG nova.virt.hardware [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1156.062959] env[62460]: DEBUG nova.virt.hardware [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1156.063113] env[62460]: DEBUG nova.virt.hardware [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1156.063308] env[62460]: DEBUG nova.virt.hardware [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1156.063500] env[62460]: DEBUG nova.virt.hardware [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1156.063617] env[62460]: DEBUG nova.virt.hardware [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1156.063831] env[62460]: DEBUG nova.virt.hardware [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1156.064008] env[62460]: DEBUG nova.virt.hardware [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1156.064192] env[62460]: DEBUG nova.virt.hardware [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1156.064364] env[62460]: DEBUG nova.virt.hardware [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1156.064545] env[62460]: DEBUG nova.virt.hardware [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1156.069532] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-74ed1dee-9451-4e04-aff2-e0fdeb0dbb45 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.085232] env[62460]: DEBUG oslo_vmware.api [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1156.085232] env[62460]: value = "task-1314165" [ 1156.085232] env[62460]: _type = "Task" [ 1156.085232] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.094604] env[62460]: DEBUG oslo_vmware.api [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314165, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.595694] env[62460]: DEBUG oslo_vmware.api [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314165, 'name': ReconfigVM_Task, 'duration_secs': 0.164188} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.596089] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Updating instance '61f4666e-0588-4a00-a969-2349148caef2' progress to 33 {{(pid=62460) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1157.102585] env[62460]: DEBUG nova.virt.hardware [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1157.102871] env[62460]: DEBUG nova.virt.hardware [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1157.102995] env[62460]: DEBUG nova.virt.hardware [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1157.103201] env[62460]: DEBUG nova.virt.hardware [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1157.103358] env[62460]: DEBUG nova.virt.hardware [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1157.103532] env[62460]: DEBUG nova.virt.hardware [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1157.103720] env[62460]: DEBUG nova.virt.hardware [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1157.104011] env[62460]: DEBUG nova.virt.hardware [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1157.104095] env[62460]: DEBUG nova.virt.hardware [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1157.104241] env[62460]: DEBUG nova.virt.hardware [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1157.104418] env[62460]: DEBUG nova.virt.hardware [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1157.109651] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Reconfiguring VM instance instance-00000063 to detach disk 2000 {{(pid=62460) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1157.109908] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a45fcc95-735d-4e33-8f21-d44c37b13f84 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.127954] env[62460]: DEBUG oslo_vmware.api [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1157.127954] env[62460]: value = "task-1314166" [ 1157.127954] env[62460]: _type = "Task" [ 1157.127954] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.135309] env[62460]: DEBUG oslo_vmware.api [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314166, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.637012] env[62460]: DEBUG oslo_vmware.api [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314166, 'name': ReconfigVM_Task, 'duration_secs': 0.151183} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.637310] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Reconfigured VM instance instance-00000063 to detach disk 2000 {{(pid=62460) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1157.638059] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc5b4c80-bcfe-456c-8a05-2a91f0e5cdb1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.659828] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] 61f4666e-0588-4a00-a969-2349148caef2/61f4666e-0588-4a00-a969-2349148caef2.vmdk or device None with type thin {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1157.660068] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c29129f4-a1c6-4e42-b43d-57b5d7955538 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.677718] env[62460]: DEBUG oslo_vmware.api [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1157.677718] env[62460]: value = "task-1314167" [ 1157.677718] env[62460]: _type = "Task" [ 1157.677718] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.684719] env[62460]: DEBUG oslo_vmware.api [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314167, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.187516] env[62460]: DEBUG oslo_vmware.api [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314167, 'name': ReconfigVM_Task, 'duration_secs': 0.247335} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.187805] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Reconfigured VM instance instance-00000063 to attach disk [datastore1] 61f4666e-0588-4a00-a969-2349148caef2/61f4666e-0588-4a00-a969-2349148caef2.vmdk or device None with type thin {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1158.188093] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Updating instance '61f4666e-0588-4a00-a969-2349148caef2' progress to 50 {{(pid=62460) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1158.695313] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4efc823a-1e64-4b3a-b0ed-6a5094152485 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.713785] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75bc010a-bd70-43b2-b76f-1533572a64dd {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.730461] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Updating instance '61f4666e-0588-4a00-a969-2349148caef2' progress to 67 {{(pid=62460) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1159.269953] env[62460]: DEBUG nova.network.neutron [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Port 345c87ec-2bb1-420f-bc8c-845a795baad5 binding to destination host cpu-1 is already ACTIVE {{(pid=62460) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1160.292203] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "61f4666e-0588-4a00-a969-2349148caef2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1160.292557] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "61f4666e-0588-4a00-a969-2349148caef2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1160.292593] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "61f4666e-0588-4a00-a969-2349148caef2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1161.328166] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "refresh_cache-61f4666e-0588-4a00-a969-2349148caef2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1161.328423] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquired lock "refresh_cache-61f4666e-0588-4a00-a969-2349148caef2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1161.328611] env[62460]: DEBUG nova.network.neutron [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1162.074369] env[62460]: DEBUG nova.network.neutron [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Updating instance_info_cache with network_info: [{"id": "345c87ec-2bb1-420f-bc8c-845a795baad5", "address": "fa:16:3e:43:36:b0", "network": {"id": "1dfc278d-e89e-4caf-9cfb-c302106cf974", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-868970047-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eee8bf23869e4680aec16e37f972175f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap345c87ec-2b", "ovs_interfaceid": "345c87ec-2bb1-420f-bc8c-845a795baad5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1162.577105] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Releasing lock "refresh_cache-61f4666e-0588-4a00-a969-2349148caef2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1163.101902] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f7ca002-d360-4e48-b787-422d26fa0956 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.120378] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81f0dde1-cb47-494a-b6a8-7f9bb6d0a2b7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.126974] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Updating instance '61f4666e-0588-4a00-a969-2349148caef2' progress to 83 {{(pid=62460) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1163.633512] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1163.633882] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7558cb47-8841-47d6-8754-e0982cc031e5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.640822] env[62460]: DEBUG oslo_vmware.api [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1163.640822] env[62460]: value = "task-1314168" [ 1163.640822] env[62460]: _type = "Task" [ 1163.640822] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.648331] env[62460]: DEBUG oslo_vmware.api [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314168, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.150914] env[62460]: DEBUG oslo_vmware.api [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314168, 'name': PowerOnVM_Task, 'duration_secs': 0.349799} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.151253] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1164.151490] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2ddf554c-befe-4534-b726-dd05007eb73b tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Updating instance '61f4666e-0588-4a00-a969-2349148caef2' progress to 100 {{(pid=62460) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1165.913014] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6bb5a032-bb3b-45de-bd4a-f747a0c43d91 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "61f4666e-0588-4a00-a969-2349148caef2" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1165.913470] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6bb5a032-bb3b-45de-bd4a-f747a0c43d91 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "61f4666e-0588-4a00-a969-2349148caef2" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1165.913584] env[62460]: DEBUG nova.compute.manager [None req-6bb5a032-bb3b-45de-bd4a-f747a0c43d91 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Going to confirm migration 4 {{(pid=62460) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1166.482978] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6bb5a032-bb3b-45de-bd4a-f747a0c43d91 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "refresh_cache-61f4666e-0588-4a00-a969-2349148caef2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1166.483217] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6bb5a032-bb3b-45de-bd4a-f747a0c43d91 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquired lock "refresh_cache-61f4666e-0588-4a00-a969-2349148caef2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1166.483411] env[62460]: DEBUG nova.network.neutron [None req-6bb5a032-bb3b-45de-bd4a-f747a0c43d91 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1166.483606] env[62460]: DEBUG nova.objects.instance [None req-6bb5a032-bb3b-45de-bd4a-f747a0c43d91 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lazy-loading 'info_cache' on Instance uuid 61f4666e-0588-4a00-a969-2349148caef2 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1167.684663] env[62460]: DEBUG nova.network.neutron [None req-6bb5a032-bb3b-45de-bd4a-f747a0c43d91 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Updating instance_info_cache with network_info: [{"id": "345c87ec-2bb1-420f-bc8c-845a795baad5", "address": "fa:16:3e:43:36:b0", "network": {"id": "1dfc278d-e89e-4caf-9cfb-c302106cf974", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-868970047-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eee8bf23869e4680aec16e37f972175f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap345c87ec-2b", "ovs_interfaceid": "345c87ec-2bb1-420f-bc8c-845a795baad5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1168.187926] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6bb5a032-bb3b-45de-bd4a-f747a0c43d91 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Releasing lock "refresh_cache-61f4666e-0588-4a00-a969-2349148caef2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1168.188201] env[62460]: DEBUG nova.objects.instance [None req-6bb5a032-bb3b-45de-bd4a-f747a0c43d91 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lazy-loading 'migration_context' on Instance uuid 61f4666e-0588-4a00-a969-2349148caef2 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1168.691101] env[62460]: DEBUG nova.objects.base [None req-6bb5a032-bb3b-45de-bd4a-f747a0c43d91 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Object Instance<61f4666e-0588-4a00-a969-2349148caef2> lazy-loaded attributes: info_cache,migration_context {{(pid=62460) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1168.692100] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40eac4d5-469f-4279-a35f-b66e2cf26b85 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.711258] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4948ea24-e2ba-4823-9c41-8b1c1db2b7d5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.716581] env[62460]: DEBUG oslo_vmware.api [None req-6bb5a032-bb3b-45de-bd4a-f747a0c43d91 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1168.716581] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52cf1850-c595-3fbf-d2af-1c71f7f277f5" [ 1168.716581] env[62460]: _type = "Task" [ 1168.716581] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.724213] env[62460]: DEBUG oslo_vmware.api [None req-6bb5a032-bb3b-45de-bd4a-f747a0c43d91 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52cf1850-c595-3fbf-d2af-1c71f7f277f5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.227674] env[62460]: DEBUG oslo_vmware.api [None req-6bb5a032-bb3b-45de-bd4a-f747a0c43d91 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52cf1850-c595-3fbf-d2af-1c71f7f277f5, 'name': SearchDatastore_Task, 'duration_secs': 0.010199} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.227956] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6bb5a032-bb3b-45de-bd4a-f747a0c43d91 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1169.228216] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6bb5a032-bb3b-45de-bd4a-f747a0c43d91 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1169.771910] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c72fdf7-5724-4841-bf63-9643f82a0062 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.779299] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c3c1cd9-667e-49b3-b941-b1f4e63b0924 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.808996] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d755425-d216-47b7-a6ca-230eef408e92 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.815576] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f704fb6-30a0-4a27-af47-10ad19f6aa55 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.828377] env[62460]: DEBUG nova.compute.provider_tree [None req-6bb5a032-bb3b-45de-bd4a-f747a0c43d91 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1170.331590] env[62460]: DEBUG nova.scheduler.client.report [None req-6bb5a032-bb3b-45de-bd4a-f747a0c43d91 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1171.342027] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6bb5a032-bb3b-45de-bd4a-f747a0c43d91 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.114s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1171.893241] env[62460]: INFO nova.scheduler.client.report [None req-6bb5a032-bb3b-45de-bd4a-f747a0c43d91 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Deleted allocation for migration e7a3c104-1792-4b07-9214-3984fb4e1c59 [ 1172.398246] env[62460]: DEBUG oslo_concurrency.lockutils [None req-6bb5a032-bb3b-45de-bd4a-f747a0c43d91 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "61f4666e-0588-4a00-a969-2349148caef2" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.485s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1172.868521] env[62460]: DEBUG oslo_concurrency.lockutils [None req-bd127d93-b7cf-48e7-a61e-fa48d925e8d4 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "61f4666e-0588-4a00-a969-2349148caef2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1172.868788] env[62460]: DEBUG oslo_concurrency.lockutils [None req-bd127d93-b7cf-48e7-a61e-fa48d925e8d4 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "61f4666e-0588-4a00-a969-2349148caef2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1172.869018] env[62460]: DEBUG oslo_concurrency.lockutils [None req-bd127d93-b7cf-48e7-a61e-fa48d925e8d4 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "61f4666e-0588-4a00-a969-2349148caef2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1172.869218] env[62460]: DEBUG oslo_concurrency.lockutils [None req-bd127d93-b7cf-48e7-a61e-fa48d925e8d4 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "61f4666e-0588-4a00-a969-2349148caef2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1172.869400] env[62460]: DEBUG oslo_concurrency.lockutils [None req-bd127d93-b7cf-48e7-a61e-fa48d925e8d4 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "61f4666e-0588-4a00-a969-2349148caef2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1172.871598] env[62460]: INFO nova.compute.manager [None req-bd127d93-b7cf-48e7-a61e-fa48d925e8d4 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Terminating instance [ 1172.873388] env[62460]: DEBUG nova.compute.manager [None req-bd127d93-b7cf-48e7-a61e-fa48d925e8d4 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1172.873590] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-bd127d93-b7cf-48e7-a61e-fa48d925e8d4 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1172.874440] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a17960ff-1c2c-49ef-b049-32bed1943488 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.882358] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd127d93-b7cf-48e7-a61e-fa48d925e8d4 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1172.882873] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-59daf17e-b145-4e87-9795-d895714401f8 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.889114] env[62460]: DEBUG oslo_vmware.api [None req-bd127d93-b7cf-48e7-a61e-fa48d925e8d4 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1172.889114] env[62460]: value = "task-1314169" [ 1172.889114] env[62460]: _type = "Task" [ 1172.889114] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.896252] env[62460]: DEBUG oslo_vmware.api [None req-bd127d93-b7cf-48e7-a61e-fa48d925e8d4 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314169, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.399103] env[62460]: DEBUG oslo_vmware.api [None req-bd127d93-b7cf-48e7-a61e-fa48d925e8d4 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314169, 'name': PowerOffVM_Task, 'duration_secs': 0.190684} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.399468] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd127d93-b7cf-48e7-a61e-fa48d925e8d4 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1173.399570] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-bd127d93-b7cf-48e7-a61e-fa48d925e8d4 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1173.399819] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ee28ba50-a886-4316-b596-e080b1273424 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.455928] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-bd127d93-b7cf-48e7-a61e-fa48d925e8d4 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1173.456173] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-bd127d93-b7cf-48e7-a61e-fa48d925e8d4 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Deleting contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1173.456362] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd127d93-b7cf-48e7-a61e-fa48d925e8d4 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Deleting the datastore file [datastore1] 61f4666e-0588-4a00-a969-2349148caef2 {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1173.456617] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bf21f710-461d-4204-8cd0-f8158a6090ea {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.462793] env[62460]: DEBUG oslo_vmware.api [None req-bd127d93-b7cf-48e7-a61e-fa48d925e8d4 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1173.462793] env[62460]: value = "task-1314171" [ 1173.462793] env[62460]: _type = "Task" [ 1173.462793] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.470956] env[62460]: DEBUG oslo_vmware.api [None req-bd127d93-b7cf-48e7-a61e-fa48d925e8d4 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314171, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.972881] env[62460]: DEBUG oslo_vmware.api [None req-bd127d93-b7cf-48e7-a61e-fa48d925e8d4 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314171, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.135212} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.973869] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd127d93-b7cf-48e7-a61e-fa48d925e8d4 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1173.973869] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-bd127d93-b7cf-48e7-a61e-fa48d925e8d4 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Deleted contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1173.973869] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-bd127d93-b7cf-48e7-a61e-fa48d925e8d4 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1173.974050] env[62460]: INFO nova.compute.manager [None req-bd127d93-b7cf-48e7-a61e-fa48d925e8d4 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1173.974247] env[62460]: DEBUG oslo.service.loopingcall [None req-bd127d93-b7cf-48e7-a61e-fa48d925e8d4 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1173.974453] env[62460]: DEBUG nova.compute.manager [-] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1173.974550] env[62460]: DEBUG nova.network.neutron [-] [instance: 61f4666e-0588-4a00-a969-2349148caef2] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1174.396837] env[62460]: DEBUG nova.compute.manager [req-0caf6577-ba98-48e9-b76d-c22d54305425 req-1a51d9fd-e7b0-4505-a4a5-e4928ec3cde2 service nova] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Received event network-vif-deleted-345c87ec-2bb1-420f-bc8c-845a795baad5 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1174.397064] env[62460]: INFO nova.compute.manager [req-0caf6577-ba98-48e9-b76d-c22d54305425 req-1a51d9fd-e7b0-4505-a4a5-e4928ec3cde2 service nova] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Neutron deleted interface 345c87ec-2bb1-420f-bc8c-845a795baad5; detaching it from the instance and deleting it from the info cache [ 1174.397247] env[62460]: DEBUG nova.network.neutron [req-0caf6577-ba98-48e9-b76d-c22d54305425 req-1a51d9fd-e7b0-4505-a4a5-e4928ec3cde2 service nova] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1174.873838] env[62460]: DEBUG nova.network.neutron [-] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1174.899206] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6bf5058b-b8ff-42b5-8bc9-78b9ac95cfbf {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.909287] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-165bbb0f-d01e-467b-ad4c-f9f9700cd537 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.932121] env[62460]: DEBUG nova.compute.manager [req-0caf6577-ba98-48e9-b76d-c22d54305425 req-1a51d9fd-e7b0-4505-a4a5-e4928ec3cde2 service nova] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Detach interface failed, port_id=345c87ec-2bb1-420f-bc8c-845a795baad5, reason: Instance 61f4666e-0588-4a00-a969-2349148caef2 could not be found. {{(pid=62460) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1175.376403] env[62460]: INFO nova.compute.manager [-] [instance: 61f4666e-0588-4a00-a969-2349148caef2] Took 1.40 seconds to deallocate network for instance. [ 1175.882609] env[62460]: DEBUG oslo_concurrency.lockutils [None req-bd127d93-b7cf-48e7-a61e-fa48d925e8d4 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1175.882986] env[62460]: DEBUG oslo_concurrency.lockutils [None req-bd127d93-b7cf-48e7-a61e-fa48d925e8d4 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1175.883106] env[62460]: DEBUG oslo_concurrency.lockutils [None req-bd127d93-b7cf-48e7-a61e-fa48d925e8d4 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1175.903599] env[62460]: INFO nova.scheduler.client.report [None req-bd127d93-b7cf-48e7-a61e-fa48d925e8d4 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Deleted allocations for instance 61f4666e-0588-4a00-a969-2349148caef2 [ 1176.411925] env[62460]: DEBUG oslo_concurrency.lockutils [None req-bd127d93-b7cf-48e7-a61e-fa48d925e8d4 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "61f4666e-0588-4a00-a969-2349148caef2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.543s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1177.785951] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "bc47bebb-fabb-4d6f-ac7e-8fa674e408a2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1177.786307] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "bc47bebb-fabb-4d6f-ac7e-8fa674e408a2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1178.288901] env[62460]: DEBUG nova.compute.manager [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1178.809428] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1178.809744] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1178.811330] env[62460]: INFO nova.compute.claims [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1179.846156] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fe376d0-0092-425b-bace-e1d12904be53 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.853669] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79f5eb0a-c4da-46f4-95b8-469c68eb0676 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.883242] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a7e7214-d268-4edc-a3b2-c9a2a352767b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.889736] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-633d443b-1e32-4def-9631-45b8741f5202 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.902174] env[62460]: DEBUG nova.compute.provider_tree [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1180.405234] env[62460]: DEBUG nova.scheduler.client.report [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1180.910839] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.101s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1180.911353] env[62460]: DEBUG nova.compute.manager [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1181.416662] env[62460]: DEBUG nova.compute.utils [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1181.418146] env[62460]: DEBUG nova.compute.manager [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1181.418323] env[62460]: DEBUG nova.network.neutron [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1181.465802] env[62460]: DEBUG nova.policy [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a73816ab645e4cd49fabdbc4eeb31f97', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eee8bf23869e4680aec16e37f972175f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 1181.723345] env[62460]: DEBUG nova.network.neutron [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Successfully created port: 998ec738-8d95-4efa-bef5-06dd88f7e319 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1181.922130] env[62460]: DEBUG nova.compute.manager [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1182.933062] env[62460]: DEBUG nova.compute.manager [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1182.957714] env[62460]: DEBUG nova.virt.hardware [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1182.957985] env[62460]: DEBUG nova.virt.hardware [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1182.958167] env[62460]: DEBUG nova.virt.hardware [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1182.958366] env[62460]: DEBUG nova.virt.hardware [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1182.958517] env[62460]: DEBUG nova.virt.hardware [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1182.958671] env[62460]: DEBUG nova.virt.hardware [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1182.958887] env[62460]: DEBUG nova.virt.hardware [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1182.959064] env[62460]: DEBUG nova.virt.hardware [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1182.959242] env[62460]: DEBUG nova.virt.hardware [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1182.959411] env[62460]: DEBUG nova.virt.hardware [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1182.959591] env[62460]: DEBUG nova.virt.hardware [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1182.960506] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72b3643e-f1d6-4b32-bc62-5171acf6d1ee {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.968281] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65ae3bdb-84f8-472c-8d08-fafb6f10ea15 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.062842] env[62460]: DEBUG nova.compute.manager [req-d043de34-f17a-4b05-882f-97da6ac6ba6e req-508e2206-8792-4262-b9ab-4dd6c195fef2 service nova] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Received event network-vif-plugged-998ec738-8d95-4efa-bef5-06dd88f7e319 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1183.063000] env[62460]: DEBUG oslo_concurrency.lockutils [req-d043de34-f17a-4b05-882f-97da6ac6ba6e req-508e2206-8792-4262-b9ab-4dd6c195fef2 service nova] Acquiring lock "bc47bebb-fabb-4d6f-ac7e-8fa674e408a2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1183.063237] env[62460]: DEBUG oslo_concurrency.lockutils [req-d043de34-f17a-4b05-882f-97da6ac6ba6e req-508e2206-8792-4262-b9ab-4dd6c195fef2 service nova] Lock "bc47bebb-fabb-4d6f-ac7e-8fa674e408a2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1183.063417] env[62460]: DEBUG oslo_concurrency.lockutils [req-d043de34-f17a-4b05-882f-97da6ac6ba6e req-508e2206-8792-4262-b9ab-4dd6c195fef2 service nova] Lock "bc47bebb-fabb-4d6f-ac7e-8fa674e408a2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1183.063669] env[62460]: DEBUG nova.compute.manager [req-d043de34-f17a-4b05-882f-97da6ac6ba6e req-508e2206-8792-4262-b9ab-4dd6c195fef2 service nova] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] No waiting events found dispatching network-vif-plugged-998ec738-8d95-4efa-bef5-06dd88f7e319 {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1183.063760] env[62460]: WARNING nova.compute.manager [req-d043de34-f17a-4b05-882f-97da6ac6ba6e req-508e2206-8792-4262-b9ab-4dd6c195fef2 service nova] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Received unexpected event network-vif-plugged-998ec738-8d95-4efa-bef5-06dd88f7e319 for instance with vm_state building and task_state spawning. [ 1183.629247] env[62460]: DEBUG nova.network.neutron [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Successfully updated port: 998ec738-8d95-4efa-bef5-06dd88f7e319 {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1183.651252] env[62460]: DEBUG nova.compute.manager [req-fa8a3f59-c77e-4b49-b8f8-2438369d4ed8 req-cf504974-78b5-4041-80f5-f1e98ef3b418 service nova] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Received event network-changed-998ec738-8d95-4efa-bef5-06dd88f7e319 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1183.651466] env[62460]: DEBUG nova.compute.manager [req-fa8a3f59-c77e-4b49-b8f8-2438369d4ed8 req-cf504974-78b5-4041-80f5-f1e98ef3b418 service nova] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Refreshing instance network info cache due to event network-changed-998ec738-8d95-4efa-bef5-06dd88f7e319. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1183.651695] env[62460]: DEBUG oslo_concurrency.lockutils [req-fa8a3f59-c77e-4b49-b8f8-2438369d4ed8 req-cf504974-78b5-4041-80f5-f1e98ef3b418 service nova] Acquiring lock "refresh_cache-bc47bebb-fabb-4d6f-ac7e-8fa674e408a2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1183.651843] env[62460]: DEBUG oslo_concurrency.lockutils [req-fa8a3f59-c77e-4b49-b8f8-2438369d4ed8 req-cf504974-78b5-4041-80f5-f1e98ef3b418 service nova] Acquired lock "refresh_cache-bc47bebb-fabb-4d6f-ac7e-8fa674e408a2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1183.652015] env[62460]: DEBUG nova.network.neutron [req-fa8a3f59-c77e-4b49-b8f8-2438369d4ed8 req-cf504974-78b5-4041-80f5-f1e98ef3b418 service nova] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Refreshing network info cache for port 998ec738-8d95-4efa-bef5-06dd88f7e319 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1184.131548] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "refresh_cache-bc47bebb-fabb-4d6f-ac7e-8fa674e408a2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1184.183290] env[62460]: DEBUG nova.network.neutron [req-fa8a3f59-c77e-4b49-b8f8-2438369d4ed8 req-cf504974-78b5-4041-80f5-f1e98ef3b418 service nova] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1184.250991] env[62460]: DEBUG nova.network.neutron [req-fa8a3f59-c77e-4b49-b8f8-2438369d4ed8 req-cf504974-78b5-4041-80f5-f1e98ef3b418 service nova] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1184.753672] env[62460]: DEBUG oslo_concurrency.lockutils [req-fa8a3f59-c77e-4b49-b8f8-2438369d4ed8 req-cf504974-78b5-4041-80f5-f1e98ef3b418 service nova] Releasing lock "refresh_cache-bc47bebb-fabb-4d6f-ac7e-8fa674e408a2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1184.754175] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquired lock "refresh_cache-bc47bebb-fabb-4d6f-ac7e-8fa674e408a2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1184.754259] env[62460]: DEBUG nova.network.neutron [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1185.283729] env[62460]: DEBUG nova.network.neutron [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1185.402030] env[62460]: DEBUG nova.network.neutron [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Updating instance_info_cache with network_info: [{"id": "998ec738-8d95-4efa-bef5-06dd88f7e319", "address": "fa:16:3e:e7:79:65", "network": {"id": "1dfc278d-e89e-4caf-9cfb-c302106cf974", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-868970047-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eee8bf23869e4680aec16e37f972175f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap998ec738-8d", "ovs_interfaceid": "998ec738-8d95-4efa-bef5-06dd88f7e319", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1185.904439] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Releasing lock "refresh_cache-bc47bebb-fabb-4d6f-ac7e-8fa674e408a2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1185.904788] env[62460]: DEBUG nova.compute.manager [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Instance network_info: |[{"id": "998ec738-8d95-4efa-bef5-06dd88f7e319", "address": "fa:16:3e:e7:79:65", "network": {"id": "1dfc278d-e89e-4caf-9cfb-c302106cf974", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-868970047-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eee8bf23869e4680aec16e37f972175f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap998ec738-8d", "ovs_interfaceid": "998ec738-8d95-4efa-bef5-06dd88f7e319", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1185.905250] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e7:79:65', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '418ddd3d-5f64-407e-8e0c-c8b81639bee9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '998ec738-8d95-4efa-bef5-06dd88f7e319', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1185.912839] env[62460]: DEBUG oslo.service.loopingcall [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1185.913069] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1185.913302] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b98b4f0a-9451-48e4-aee9-d8dd66cb3ffd {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.934278] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1185.934278] env[62460]: value = "task-1314172" [ 1185.934278] env[62460]: _type = "Task" [ 1185.934278] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.941554] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1314172, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.444497] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1314172, 'name': CreateVM_Task, 'duration_secs': 0.281139} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.444848] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1186.445328] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1186.445500] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1186.445839] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1186.446108] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1d16995a-aa01-473c-a440-e8c53227606d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.450097] env[62460]: DEBUG oslo_vmware.api [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1186.450097] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52d98e17-d2e9-95cb-63de-3562bd95fe2f" [ 1186.450097] env[62460]: _type = "Task" [ 1186.450097] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.456946] env[62460]: DEBUG oslo_vmware.api [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52d98e17-d2e9-95cb-63de-3562bd95fe2f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.960644] env[62460]: DEBUG oslo_vmware.api [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52d98e17-d2e9-95cb-63de-3562bd95fe2f, 'name': SearchDatastore_Task, 'duration_secs': 0.009987} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.961009] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1186.961306] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1186.961600] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1186.961796] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1186.962024] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1186.962328] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-459e00fe-f7d5-4d9b-8e8a-aba3d0c2fe00 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.969954] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1186.970185] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1186.970943] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1199a44a-f2ad-4994-965b-516543ca9f93 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.975891] env[62460]: DEBUG oslo_vmware.api [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1186.975891] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52c9ba56-f028-deb5-a4c3-722790a9b1ea" [ 1186.975891] env[62460]: _type = "Task" [ 1186.975891] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.983099] env[62460]: DEBUG oslo_vmware.api [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52c9ba56-f028-deb5-a4c3-722790a9b1ea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.487469] env[62460]: DEBUG oslo_vmware.api [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52c9ba56-f028-deb5-a4c3-722790a9b1ea, 'name': SearchDatastore_Task, 'duration_secs': 0.0079} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.488218] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c84d330e-5c28-4091-9171-3616b8bd2d14 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.493327] env[62460]: DEBUG oslo_vmware.api [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1187.493327] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]5297143e-2804-2967-1da6-bb8460aa3642" [ 1187.493327] env[62460]: _type = "Task" [ 1187.493327] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.501268] env[62460]: DEBUG oslo_vmware.api [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5297143e-2804-2967-1da6-bb8460aa3642, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.004082] env[62460]: DEBUG oslo_vmware.api [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5297143e-2804-2967-1da6-bb8460aa3642, 'name': SearchDatastore_Task, 'duration_secs': 0.009398} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.004082] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1188.004082] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] bc47bebb-fabb-4d6f-ac7e-8fa674e408a2/bc47bebb-fabb-4d6f-ac7e-8fa674e408a2.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1188.004304] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1fae5520-e0aa-4bc1-9098-2a7e412371f7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.010330] env[62460]: DEBUG oslo_vmware.api [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1188.010330] env[62460]: value = "task-1314173" [ 1188.010330] env[62460]: _type = "Task" [ 1188.010330] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.520397] env[62460]: DEBUG oslo_vmware.api [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314173, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.021334] env[62460]: DEBUG oslo_vmware.api [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314173, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.545884} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.021636] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] bc47bebb-fabb-4d6f-ac7e-8fa674e408a2/bc47bebb-fabb-4d6f-ac7e-8fa674e408a2.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1189.021862] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1189.022141] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7f762fb1-a673-457d-b03a-ed1ac29aa1f9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.028603] env[62460]: DEBUG oslo_vmware.api [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1189.028603] env[62460]: value = "task-1314174" [ 1189.028603] env[62460]: _type = "Task" [ 1189.028603] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.035837] env[62460]: DEBUG oslo_vmware.api [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314174, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.537521] env[62460]: DEBUG oslo_vmware.api [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314174, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059991} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.537828] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1189.538521] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8de7ba35-a388-4da0-a46a-ed808bf5bda9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.559255] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] bc47bebb-fabb-4d6f-ac7e-8fa674e408a2/bc47bebb-fabb-4d6f-ac7e-8fa674e408a2.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1189.559507] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b513b0f3-46bc-4800-92ff-68b7930c9f26 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.577514] env[62460]: DEBUG oslo_vmware.api [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1189.577514] env[62460]: value = "task-1314175" [ 1189.577514] env[62460]: _type = "Task" [ 1189.577514] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.584705] env[62460]: DEBUG oslo_vmware.api [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314175, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.087547] env[62460]: DEBUG oslo_vmware.api [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314175, 'name': ReconfigVM_Task, 'duration_secs': 0.265198} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.087791] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Reconfigured VM instance instance-0000006b to attach disk [datastore1] bc47bebb-fabb-4d6f-ac7e-8fa674e408a2/bc47bebb-fabb-4d6f-ac7e-8fa674e408a2.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1190.088451] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1e2cfc88-2a7f-452a-95bc-897de38346a2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.094382] env[62460]: DEBUG oslo_vmware.api [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1190.094382] env[62460]: value = "task-1314176" [ 1190.094382] env[62460]: _type = "Task" [ 1190.094382] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1190.101272] env[62460]: DEBUG oslo_vmware.api [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314176, 'name': Rename_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.604499] env[62460]: DEBUG oslo_vmware.api [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314176, 'name': Rename_Task, 'duration_secs': 0.134268} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.604839] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1190.605017] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e6658cbd-1e15-44ff-a771-18760f1e4435 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.611726] env[62460]: DEBUG oslo_vmware.api [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1190.611726] env[62460]: value = "task-1314177" [ 1190.611726] env[62460]: _type = "Task" [ 1190.611726] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1190.618766] env[62460]: DEBUG oslo_vmware.api [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314177, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.124696] env[62460]: DEBUG oslo_vmware.api [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314177, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.622142] env[62460]: DEBUG oslo_vmware.api [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314177, 'name': PowerOnVM_Task, 'duration_secs': 0.542276} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1191.622453] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1191.622615] env[62460]: INFO nova.compute.manager [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Took 8.69 seconds to spawn the instance on the hypervisor. [ 1191.622825] env[62460]: DEBUG nova.compute.manager [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1191.623573] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e82a3360-fa2c-4d46-8a7f-e5bc79633079 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.142897] env[62460]: INFO nova.compute.manager [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Took 13.35 seconds to build instance. [ 1192.645634] env[62460]: DEBUG oslo_concurrency.lockutils [None req-2c3c04d5-fd88-411c-94c8-4c59800ba880 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "bc47bebb-fabb-4d6f-ac7e-8fa674e408a2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.859s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1193.183790] env[62460]: DEBUG nova.compute.manager [req-bca54e15-2448-469a-bf95-da0f31c9d09a req-b7d26cab-c6c3-4d1c-975d-8328fcb35724 service nova] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Received event network-changed-998ec738-8d95-4efa-bef5-06dd88f7e319 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1193.183998] env[62460]: DEBUG nova.compute.manager [req-bca54e15-2448-469a-bf95-da0f31c9d09a req-b7d26cab-c6c3-4d1c-975d-8328fcb35724 service nova] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Refreshing instance network info cache due to event network-changed-998ec738-8d95-4efa-bef5-06dd88f7e319. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1193.184417] env[62460]: DEBUG oslo_concurrency.lockutils [req-bca54e15-2448-469a-bf95-da0f31c9d09a req-b7d26cab-c6c3-4d1c-975d-8328fcb35724 service nova] Acquiring lock "refresh_cache-bc47bebb-fabb-4d6f-ac7e-8fa674e408a2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1193.184573] env[62460]: DEBUG oslo_concurrency.lockutils [req-bca54e15-2448-469a-bf95-da0f31c9d09a req-b7d26cab-c6c3-4d1c-975d-8328fcb35724 service nova] Acquired lock "refresh_cache-bc47bebb-fabb-4d6f-ac7e-8fa674e408a2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1193.184822] env[62460]: DEBUG nova.network.neutron [req-bca54e15-2448-469a-bf95-da0f31c9d09a req-b7d26cab-c6c3-4d1c-975d-8328fcb35724 service nova] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Refreshing network info cache for port 998ec738-8d95-4efa-bef5-06dd88f7e319 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1193.902656] env[62460]: DEBUG nova.network.neutron [req-bca54e15-2448-469a-bf95-da0f31c9d09a req-b7d26cab-c6c3-4d1c-975d-8328fcb35724 service nova] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Updated VIF entry in instance network info cache for port 998ec738-8d95-4efa-bef5-06dd88f7e319. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1193.903135] env[62460]: DEBUG nova.network.neutron [req-bca54e15-2448-469a-bf95-da0f31c9d09a req-b7d26cab-c6c3-4d1c-975d-8328fcb35724 service nova] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Updating instance_info_cache with network_info: [{"id": "998ec738-8d95-4efa-bef5-06dd88f7e319", "address": "fa:16:3e:e7:79:65", "network": {"id": "1dfc278d-e89e-4caf-9cfb-c302106cf974", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-868970047-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eee8bf23869e4680aec16e37f972175f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap998ec738-8d", "ovs_interfaceid": "998ec738-8d95-4efa-bef5-06dd88f7e319", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1194.405833] env[62460]: DEBUG oslo_concurrency.lockutils [req-bca54e15-2448-469a-bf95-da0f31c9d09a req-b7d26cab-c6c3-4d1c-975d-8328fcb35724 service nova] Releasing lock "refresh_cache-bc47bebb-fabb-4d6f-ac7e-8fa674e408a2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1196.592985] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1196.593369] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Starting heal instance info cache {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1197.095859] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Didn't find any instances for network info cache update. {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1197.096095] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1197.096260] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1197.096422] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1197.096575] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1197.096710] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62460) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1197.868989] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1197.869420] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1198.868993] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1200.869024] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1201.372495] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1201.372823] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1201.373018] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1201.373183] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62460) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1201.374343] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a521048a-5dbb-450f-8ce0-fcefa813ee56 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.382389] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f454ee58-7d8d-4c05-bc48-1b51472b6550 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.395815] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a1cfdb6-29e6-4e48-a185-1abb8d530682 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.401866] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-778b784d-25a8-4701-ad8c-38452303c228 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.429020] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181137MB free_disk=135GB free_vcpus=48 pci_devices=None {{(pid=62460) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1201.429171] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1201.429361] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1202.452247] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance bc47bebb-fabb-4d6f-ac7e-8fa674e408a2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1202.452495] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=62460) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1202.452681] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=62460) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1202.478603] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33f646d8-7676-4268-8ba0-6119df1eee8e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.486493] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e0bed15-4adb-453c-ba0b-5c64ef9344e5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.514784] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc821afd-e849-4286-a69b-3c76aff4b60e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.521280] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-855d3072-efcd-4bf4-8c03-1d507c1394a4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.534801] env[62460]: DEBUG nova.compute.provider_tree [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1203.037731] env[62460]: DEBUG nova.scheduler.client.report [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1203.542181] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62460) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1203.542532] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.113s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1204.537629] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1231.154902] env[62460]: DEBUG nova.compute.manager [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Stashing vm_state: active {{(pid=62460) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1231.672424] env[62460]: DEBUG oslo_concurrency.lockutils [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1231.672698] env[62460]: DEBUG oslo_concurrency.lockutils [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1232.177552] env[62460]: INFO nova.compute.claims [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1232.684145] env[62460]: INFO nova.compute.resource_tracker [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Updating resource usage from migration e4adb8a9-596a-4a99-99c8-6ff7f5041925 [ 1232.720194] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b1ea837-466c-44a5-acf9-62a731a56a9a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.727758] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec7dd3f1-4554-4d28-a8e6-823242b049ea {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.756440] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32d9166e-3b85-4c4f-b350-bcaaf3541e0b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.763338] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-361b2b35-0158-4f6e-bb4a-f29d165688a1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.775823] env[62460]: DEBUG nova.compute.provider_tree [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1233.279376] env[62460]: DEBUG nova.scheduler.client.report [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1233.784917] env[62460]: DEBUG oslo_concurrency.lockutils [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.112s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1233.785175] env[62460]: INFO nova.compute.manager [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Migrating [ 1234.299269] env[62460]: DEBUG oslo_concurrency.lockutils [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "refresh_cache-bc47bebb-fabb-4d6f-ac7e-8fa674e408a2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1234.299635] env[62460]: DEBUG oslo_concurrency.lockutils [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquired lock "refresh_cache-bc47bebb-fabb-4d6f-ac7e-8fa674e408a2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1234.299719] env[62460]: DEBUG nova.network.neutron [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1235.014461] env[62460]: DEBUG nova.network.neutron [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Updating instance_info_cache with network_info: [{"id": "998ec738-8d95-4efa-bef5-06dd88f7e319", "address": "fa:16:3e:e7:79:65", "network": {"id": "1dfc278d-e89e-4caf-9cfb-c302106cf974", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-868970047-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eee8bf23869e4680aec16e37f972175f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap998ec738-8d", "ovs_interfaceid": "998ec738-8d95-4efa-bef5-06dd88f7e319", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1235.517750] env[62460]: DEBUG oslo_concurrency.lockutils [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Releasing lock "refresh_cache-bc47bebb-fabb-4d6f-ac7e-8fa674e408a2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1237.032803] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7e5e649-e385-4fc3-a800-27f81282c6f9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.050862] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Updating instance 'bc47bebb-fabb-4d6f-ac7e-8fa674e408a2' progress to 0 {{(pid=62460) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1237.557477] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1237.557790] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f975094e-27fe-4cae-bffd-dd4974a7f5a4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.564892] env[62460]: DEBUG oslo_vmware.api [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1237.564892] env[62460]: value = "task-1314178" [ 1237.564892] env[62460]: _type = "Task" [ 1237.564892] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1237.572358] env[62460]: DEBUG oslo_vmware.api [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314178, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1238.075255] env[62460]: DEBUG oslo_vmware.api [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314178, 'name': PowerOffVM_Task, 'duration_secs': 0.166037} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1238.075586] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1238.075706] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Updating instance 'bc47bebb-fabb-4d6f-ac7e-8fa674e408a2' progress to 17 {{(pid=62460) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1238.581800] env[62460]: DEBUG nova.virt.hardware [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1238.582093] env[62460]: DEBUG nova.virt.hardware [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1238.582270] env[62460]: DEBUG nova.virt.hardware [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1238.582461] env[62460]: DEBUG nova.virt.hardware [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1238.582616] env[62460]: DEBUG nova.virt.hardware [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1238.582778] env[62460]: DEBUG nova.virt.hardware [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1238.582990] env[62460]: DEBUG nova.virt.hardware [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1238.583169] env[62460]: DEBUG nova.virt.hardware [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1238.583344] env[62460]: DEBUG nova.virt.hardware [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1238.583515] env[62460]: DEBUG nova.virt.hardware [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1238.583696] env[62460]: DEBUG nova.virt.hardware [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1238.588604] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3debda66-f0e7-4bda-8042-929461748106 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.604486] env[62460]: DEBUG oslo_vmware.api [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1238.604486] env[62460]: value = "task-1314179" [ 1238.604486] env[62460]: _type = "Task" [ 1238.604486] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1238.612375] env[62460]: DEBUG oslo_vmware.api [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314179, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1239.114591] env[62460]: DEBUG oslo_vmware.api [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314179, 'name': ReconfigVM_Task, 'duration_secs': 0.150453} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1239.115024] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Updating instance 'bc47bebb-fabb-4d6f-ac7e-8fa674e408a2' progress to 33 {{(pid=62460) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1239.621850] env[62460]: DEBUG nova.virt.hardware [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1239.622102] env[62460]: DEBUG nova.virt.hardware [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1239.622277] env[62460]: DEBUG nova.virt.hardware [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1239.622468] env[62460]: DEBUG nova.virt.hardware [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1239.622622] env[62460]: DEBUG nova.virt.hardware [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1239.622772] env[62460]: DEBUG nova.virt.hardware [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1239.622980] env[62460]: DEBUG nova.virt.hardware [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1239.623164] env[62460]: DEBUG nova.virt.hardware [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1239.623339] env[62460]: DEBUG nova.virt.hardware [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1239.623508] env[62460]: DEBUG nova.virt.hardware [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1239.623687] env[62460]: DEBUG nova.virt.hardware [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1239.628873] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Reconfiguring VM instance instance-0000006b to detach disk 2000 {{(pid=62460) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1239.629175] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fe7c5ec5-abf8-4e8d-91d5-6693452fcdea {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.646542] env[62460]: DEBUG oslo_vmware.api [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1239.646542] env[62460]: value = "task-1314180" [ 1239.646542] env[62460]: _type = "Task" [ 1239.646542] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1239.654089] env[62460]: DEBUG oslo_vmware.api [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314180, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1240.155982] env[62460]: DEBUG oslo_vmware.api [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314180, 'name': ReconfigVM_Task, 'duration_secs': 0.154548} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1240.156352] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Reconfigured VM instance instance-0000006b to detach disk 2000 {{(pid=62460) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1240.157084] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb1419e4-e5d0-44f5-b6aa-9db04a668217 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.179119] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] bc47bebb-fabb-4d6f-ac7e-8fa674e408a2/bc47bebb-fabb-4d6f-ac7e-8fa674e408a2.vmdk or device None with type thin {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1240.179333] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1f0be36e-792b-4f05-a2a9-c8991c51a9c7 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.198234] env[62460]: DEBUG oslo_vmware.api [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1240.198234] env[62460]: value = "task-1314181" [ 1240.198234] env[62460]: _type = "Task" [ 1240.198234] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1240.207783] env[62460]: DEBUG oslo_vmware.api [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314181, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1240.707533] env[62460]: DEBUG oslo_vmware.api [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314181, 'name': ReconfigVM_Task, 'duration_secs': 0.259618} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1240.707839] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Reconfigured VM instance instance-0000006b to attach disk [datastore1] bc47bebb-fabb-4d6f-ac7e-8fa674e408a2/bc47bebb-fabb-4d6f-ac7e-8fa674e408a2.vmdk or device None with type thin {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1240.708191] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Updating instance 'bc47bebb-fabb-4d6f-ac7e-8fa674e408a2' progress to 50 {{(pid=62460) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1241.214879] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1f0ed52-1e0d-4c61-8949-40d93ed6af79 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.233410] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8091f9d-c2e1-47dc-a5ce-d62c87808ebb {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.249586] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Updating instance 'bc47bebb-fabb-4d6f-ac7e-8fa674e408a2' progress to 67 {{(pid=62460) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1241.789565] env[62460]: DEBUG nova.network.neutron [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Port 998ec738-8d95-4efa-bef5-06dd88f7e319 binding to destination host cpu-1 is already ACTIVE {{(pid=62460) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1242.810736] env[62460]: DEBUG oslo_concurrency.lockutils [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "bc47bebb-fabb-4d6f-ac7e-8fa674e408a2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1242.811204] env[62460]: DEBUG oslo_concurrency.lockutils [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "bc47bebb-fabb-4d6f-ac7e-8fa674e408a2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1242.811204] env[62460]: DEBUG oslo_concurrency.lockutils [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "bc47bebb-fabb-4d6f-ac7e-8fa674e408a2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1243.846017] env[62460]: DEBUG oslo_concurrency.lockutils [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "refresh_cache-bc47bebb-fabb-4d6f-ac7e-8fa674e408a2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1243.846305] env[62460]: DEBUG oslo_concurrency.lockutils [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquired lock "refresh_cache-bc47bebb-fabb-4d6f-ac7e-8fa674e408a2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1243.846420] env[62460]: DEBUG nova.network.neutron [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1244.545654] env[62460]: DEBUG nova.network.neutron [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Updating instance_info_cache with network_info: [{"id": "998ec738-8d95-4efa-bef5-06dd88f7e319", "address": "fa:16:3e:e7:79:65", "network": {"id": "1dfc278d-e89e-4caf-9cfb-c302106cf974", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-868970047-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eee8bf23869e4680aec16e37f972175f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap998ec738-8d", "ovs_interfaceid": "998ec738-8d95-4efa-bef5-06dd88f7e319", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1245.048482] env[62460]: DEBUG oslo_concurrency.lockutils [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Releasing lock "refresh_cache-bc47bebb-fabb-4d6f-ac7e-8fa674e408a2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1245.571796] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af23c53b-a633-44ff-aeae-9bd22b8753c1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.591384] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1d6e756-e046-4821-a928-79cc72508118 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.597723] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Updating instance 'bc47bebb-fabb-4d6f-ac7e-8fa674e408a2' progress to 83 {{(pid=62460) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1246.104066] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1246.104426] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7ffbd151-d81e-4100-bac3-37dc64bead8f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.111674] env[62460]: DEBUG oslo_vmware.api [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1246.111674] env[62460]: value = "task-1314182" [ 1246.111674] env[62460]: _type = "Task" [ 1246.111674] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1246.119408] env[62460]: DEBUG oslo_vmware.api [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314182, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1246.622859] env[62460]: DEBUG oslo_vmware.api [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314182, 'name': PowerOnVM_Task, 'duration_secs': 0.351435} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1246.623137] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1246.623333] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-86b6a7cb-b732-454e-8858-89ab386f5ba6 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Updating instance 'bc47bebb-fabb-4d6f-ac7e-8fa674e408a2' progress to 100 {{(pid=62460) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1249.536139] env[62460]: DEBUG nova.network.neutron [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Port 998ec738-8d95-4efa-bef5-06dd88f7e319 binding to destination host cpu-1 is already ACTIVE {{(pid=62460) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1249.536561] env[62460]: DEBUG oslo_concurrency.lockutils [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "refresh_cache-bc47bebb-fabb-4d6f-ac7e-8fa674e408a2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1249.536787] env[62460]: DEBUG oslo_concurrency.lockutils [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquired lock "refresh_cache-bc47bebb-fabb-4d6f-ac7e-8fa674e408a2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1249.537085] env[62460]: DEBUG nova.network.neutron [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1250.290976] env[62460]: DEBUG nova.network.neutron [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Updating instance_info_cache with network_info: [{"id": "998ec738-8d95-4efa-bef5-06dd88f7e319", "address": "fa:16:3e:e7:79:65", "network": {"id": "1dfc278d-e89e-4caf-9cfb-c302106cf974", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-868970047-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eee8bf23869e4680aec16e37f972175f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap998ec738-8d", "ovs_interfaceid": "998ec738-8d95-4efa-bef5-06dd88f7e319", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1250.794061] env[62460]: DEBUG oslo_concurrency.lockutils [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Releasing lock "refresh_cache-bc47bebb-fabb-4d6f-ac7e-8fa674e408a2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1251.297821] env[62460]: DEBUG nova.compute.manager [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=62460) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1251.298094] env[62460]: DEBUG oslo_concurrency.lockutils [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1251.298369] env[62460]: DEBUG oslo_concurrency.lockutils [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1251.801773] env[62460]: DEBUG nova.objects.instance [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lazy-loading 'migration_context' on Instance uuid bc47bebb-fabb-4d6f-ac7e-8fa674e408a2 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1252.344565] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b1c4731-dacf-4bc3-b784-b7c4660167ff {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.351950] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bdb8cdd-7128-4c97-8e2f-0fefaa279f51 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.381721] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51cb5159-6078-4d9e-aab5-7dca37318d36 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.388663] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a40d5bb-ce3b-4b0e-93a1-9b6ef1e73b49 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.401245] env[62460]: DEBUG nova.compute.provider_tree [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1252.904407] env[62460]: DEBUG nova.scheduler.client.report [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1253.916076] env[62460]: DEBUG oslo_concurrency.lockutils [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.616s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1255.450371] env[62460]: INFO nova.compute.manager [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Swapping old allocation on dict_keys(['76e8f00d-65cc-4766-bb9d-5006fb51c728']) held by migration e4adb8a9-596a-4a99-99c8-6ff7f5041925 for instance [ 1255.469954] env[62460]: DEBUG nova.scheduler.client.report [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Overwriting current allocation {'allocations': {'76e8f00d-65cc-4766-bb9d-5006fb51c728': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 146}}, 'project_id': 'eee8bf23869e4680aec16e37f972175f', 'user_id': 'a73816ab645e4cd49fabdbc4eeb31f97', 'consumer_generation': 1} on consumer bc47bebb-fabb-4d6f-ac7e-8fa674e408a2 {{(pid=62460) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1255.545114] env[62460]: DEBUG oslo_concurrency.lockutils [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "refresh_cache-bc47bebb-fabb-4d6f-ac7e-8fa674e408a2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1255.545360] env[62460]: DEBUG oslo_concurrency.lockutils [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquired lock "refresh_cache-bc47bebb-fabb-4d6f-ac7e-8fa674e408a2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1255.545553] env[62460]: DEBUG nova.network.neutron [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1255.868986] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1255.869199] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Starting heal instance info cache {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1255.869295] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Rebuilding the list of instances to heal {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1256.255010] env[62460]: DEBUG nova.network.neutron [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Updating instance_info_cache with network_info: [{"id": "998ec738-8d95-4efa-bef5-06dd88f7e319", "address": "fa:16:3e:e7:79:65", "network": {"id": "1dfc278d-e89e-4caf-9cfb-c302106cf974", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-868970047-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eee8bf23869e4680aec16e37f972175f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap998ec738-8d", "ovs_interfaceid": "998ec738-8d95-4efa-bef5-06dd88f7e319", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1256.372738] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Acquiring lock "refresh_cache-bc47bebb-fabb-4d6f-ac7e-8fa674e408a2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1256.758291] env[62460]: DEBUG oslo_concurrency.lockutils [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Releasing lock "refresh_cache-bc47bebb-fabb-4d6f-ac7e-8fa674e408a2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1256.758783] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1256.759079] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Acquired lock "refresh_cache-bc47bebb-fabb-4d6f-ac7e-8fa674e408a2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1256.759233] env[62460]: DEBUG nova.network.neutron [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Forcefully refreshing network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1256.759384] env[62460]: DEBUG nova.objects.instance [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lazy-loading 'info_cache' on Instance uuid bc47bebb-fabb-4d6f-ac7e-8fa674e408a2 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1256.760773] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-61a68fd9-6947-4091-8d92-186647f1c706 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.767775] env[62460]: DEBUG oslo_vmware.api [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1256.767775] env[62460]: value = "task-1314183" [ 1256.767775] env[62460]: _type = "Task" [ 1256.767775] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1256.777336] env[62460]: DEBUG oslo_vmware.api [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314183, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1257.278308] env[62460]: DEBUG oslo_vmware.api [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314183, 'name': PowerOffVM_Task, 'duration_secs': 0.178062} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1257.278571] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1257.279233] env[62460]: DEBUG nova.virt.hardware [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1257.279463] env[62460]: DEBUG nova.virt.hardware [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1257.279627] env[62460]: DEBUG nova.virt.hardware [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1257.279815] env[62460]: DEBUG nova.virt.hardware [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1257.279968] env[62460]: DEBUG nova.virt.hardware [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1257.280140] env[62460]: DEBUG nova.virt.hardware [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1257.280349] env[62460]: DEBUG nova.virt.hardware [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1257.280515] env[62460]: DEBUG nova.virt.hardware [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1257.280688] env[62460]: DEBUG nova.virt.hardware [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1257.280850] env[62460]: DEBUG nova.virt.hardware [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1257.281087] env[62460]: DEBUG nova.virt.hardware [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1257.286113] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1135bb4c-31f3-4822-bad9-2109b10b4e3d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.301469] env[62460]: DEBUG oslo_vmware.api [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1257.301469] env[62460]: value = "task-1314184" [ 1257.301469] env[62460]: _type = "Task" [ 1257.301469] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1257.310085] env[62460]: DEBUG oslo_vmware.api [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314184, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1257.811567] env[62460]: DEBUG oslo_vmware.api [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314184, 'name': ReconfigVM_Task, 'duration_secs': 0.124743} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1257.812401] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2d419c1-3337-4ee1-8541-1889f9350f61 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.830247] env[62460]: DEBUG nova.virt.hardware [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1257.830503] env[62460]: DEBUG nova.virt.hardware [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1257.830754] env[62460]: DEBUG nova.virt.hardware [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1257.830847] env[62460]: DEBUG nova.virt.hardware [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1257.831039] env[62460]: DEBUG nova.virt.hardware [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1257.831220] env[62460]: DEBUG nova.virt.hardware [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1257.831433] env[62460]: DEBUG nova.virt.hardware [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1257.831599] env[62460]: DEBUG nova.virt.hardware [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1257.831771] env[62460]: DEBUG nova.virt.hardware [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1257.831939] env[62460]: DEBUG nova.virt.hardware [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1257.832141] env[62460]: DEBUG nova.virt.hardware [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1257.832912] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe51250d-d7ee-4b4d-99f0-9ce61438b3a9 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.838205] env[62460]: DEBUG oslo_vmware.api [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1257.838205] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]5263385f-f1f0-af70-f232-bcd0e9d1338f" [ 1257.838205] env[62460]: _type = "Task" [ 1257.838205] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1257.847434] env[62460]: DEBUG oslo_vmware.api [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5263385f-f1f0-af70-f232-bcd0e9d1338f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.347545] env[62460]: DEBUG oslo_vmware.api [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5263385f-f1f0-af70-f232-bcd0e9d1338f, 'name': SearchDatastore_Task, 'duration_secs': 0.009475} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1258.352769] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Reconfiguring VM instance instance-0000006b to detach disk 2000 {{(pid=62460) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1258.353060] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ba92aeda-51a5-432f-9218-2ebc2e9949dc {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.372187] env[62460]: DEBUG oslo_vmware.api [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1258.372187] env[62460]: value = "task-1314185" [ 1258.372187] env[62460]: _type = "Task" [ 1258.372187] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1258.379667] env[62460]: DEBUG oslo_vmware.api [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314185, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.477149] env[62460]: DEBUG nova.network.neutron [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Updating instance_info_cache with network_info: [{"id": "998ec738-8d95-4efa-bef5-06dd88f7e319", "address": "fa:16:3e:e7:79:65", "network": {"id": "1dfc278d-e89e-4caf-9cfb-c302106cf974", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-868970047-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eee8bf23869e4680aec16e37f972175f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap998ec738-8d", "ovs_interfaceid": "998ec738-8d95-4efa-bef5-06dd88f7e319", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1258.881909] env[62460]: DEBUG oslo_vmware.api [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314185, 'name': ReconfigVM_Task, 'duration_secs': 0.156253} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1258.882284] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Reconfigured VM instance instance-0000006b to detach disk 2000 {{(pid=62460) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1258.882983] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dd0296d-2b09-46a5-b53b-685f7adf5113 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.904055] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] bc47bebb-fabb-4d6f-ac7e-8fa674e408a2/bc47bebb-fabb-4d6f-ac7e-8fa674e408a2.vmdk or device None with type thin {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1258.904251] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-126546e1-0095-43fe-8d36-d6d4bad86e68 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.921835] env[62460]: DEBUG oslo_vmware.api [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1258.921835] env[62460]: value = "task-1314186" [ 1258.921835] env[62460]: _type = "Task" [ 1258.921835] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1258.933338] env[62460]: DEBUG oslo_vmware.api [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314186, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.980077] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Releasing lock "refresh_cache-bc47bebb-fabb-4d6f-ac7e-8fa674e408a2" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1258.980610] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Updated the network info_cache for instance {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1258.981792] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1258.981792] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1258.981792] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1258.981792] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1258.981792] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1258.981792] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1258.982234] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62460) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1259.431798] env[62460]: DEBUG oslo_vmware.api [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314186, 'name': ReconfigVM_Task, 'duration_secs': 0.26169} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1259.432108] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Reconfigured VM instance instance-0000006b to attach disk [datastore1] bc47bebb-fabb-4d6f-ac7e-8fa674e408a2/bc47bebb-fabb-4d6f-ac7e-8fa674e408a2.vmdk or device None with type thin {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1259.432917] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0bf0c86-f9d7-4990-aeed-24b5ed11babc {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.450057] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a284cfd5-8a9d-449b-8dcb-00f6eefaeae0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.466506] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35f42c3a-eb39-40fe-bef8-21159db70893 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.483903] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-325b195c-534f-4bc8-88a4-8dbd797dbbe3 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.489752] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1259.489966] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1cdbbecc-3a17-4e81-a773-eaaea14fa90f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.494869] env[62460]: DEBUG oslo_vmware.api [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1259.494869] env[62460]: value = "task-1314187" [ 1259.494869] env[62460]: _type = "Task" [ 1259.494869] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1259.502910] env[62460]: DEBUG oslo_vmware.api [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314187, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1260.004770] env[62460]: DEBUG oslo_vmware.api [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314187, 'name': PowerOnVM_Task, 'duration_secs': 0.350495} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1260.005157] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1261.015757] env[62460]: INFO nova.compute.manager [None req-731d7010-eb5f-4a4e-ad67-0c8608687199 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Updating instance to original state: 'active' [ 1261.869217] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1261.869448] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1261.974223] env[62460]: DEBUG oslo_concurrency.lockutils [None req-93b78cec-fea3-4267-a6a0-73ae457c669e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "bc47bebb-fabb-4d6f-ac7e-8fa674e408a2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1261.974528] env[62460]: DEBUG oslo_concurrency.lockutils [None req-93b78cec-fea3-4267-a6a0-73ae457c669e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "bc47bebb-fabb-4d6f-ac7e-8fa674e408a2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1261.974755] env[62460]: DEBUG oslo_concurrency.lockutils [None req-93b78cec-fea3-4267-a6a0-73ae457c669e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "bc47bebb-fabb-4d6f-ac7e-8fa674e408a2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1261.974947] env[62460]: DEBUG oslo_concurrency.lockutils [None req-93b78cec-fea3-4267-a6a0-73ae457c669e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "bc47bebb-fabb-4d6f-ac7e-8fa674e408a2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1261.975144] env[62460]: DEBUG oslo_concurrency.lockutils [None req-93b78cec-fea3-4267-a6a0-73ae457c669e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "bc47bebb-fabb-4d6f-ac7e-8fa674e408a2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1261.977403] env[62460]: INFO nova.compute.manager [None req-93b78cec-fea3-4267-a6a0-73ae457c669e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Terminating instance [ 1261.979259] env[62460]: DEBUG nova.compute.manager [None req-93b78cec-fea3-4267-a6a0-73ae457c669e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1261.979462] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-93b78cec-fea3-4267-a6a0-73ae457c669e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1261.980345] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1f0ea27-0550-4bab-bb3c-1b56cdea0d77 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.988354] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-93b78cec-fea3-4267-a6a0-73ae457c669e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1261.988588] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-90a2de71-af76-4ae9-9f05-1b3f291e04ea {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.993944] env[62460]: DEBUG oslo_vmware.api [None req-93b78cec-fea3-4267-a6a0-73ae457c669e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1261.993944] env[62460]: value = "task-1314188" [ 1261.993944] env[62460]: _type = "Task" [ 1261.993944] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1262.001483] env[62460]: DEBUG oslo_vmware.api [None req-93b78cec-fea3-4267-a6a0-73ae457c669e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314188, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1262.373211] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1262.373564] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1262.373676] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1262.373787] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62460) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1262.375096] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00930ce2-370a-41df-b8dd-9ed0cb1bd586 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.383316] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef217715-5bfa-474d-8ce4-bacc9bd5411e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.397277] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e78692d7-af94-4e6c-bae9-a1b87d78de4f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.403126] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d2a60fe-b90d-41ea-9782-d26a9571691a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.430347] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181490MB free_disk=135GB free_vcpus=48 pci_devices=None {{(pid=62460) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1262.430495] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1262.430684] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1262.503386] env[62460]: DEBUG oslo_vmware.api [None req-93b78cec-fea3-4267-a6a0-73ae457c669e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314188, 'name': PowerOffVM_Task, 'duration_secs': 0.141587} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1262.503588] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-93b78cec-fea3-4267-a6a0-73ae457c669e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1262.503764] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-93b78cec-fea3-4267-a6a0-73ae457c669e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1262.503996] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f3a84067-1b11-4cb6-8005-215ff6307728 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.864329] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-93b78cec-fea3-4267-a6a0-73ae457c669e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1262.864561] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-93b78cec-fea3-4267-a6a0-73ae457c669e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Deleting contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1262.864759] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-93b78cec-fea3-4267-a6a0-73ae457c669e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Deleting the datastore file [datastore1] bc47bebb-fabb-4d6f-ac7e-8fa674e408a2 {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1262.865041] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-68025ba8-4ea3-4f55-b63a-fd16254179ed {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.871655] env[62460]: DEBUG oslo_vmware.api [None req-93b78cec-fea3-4267-a6a0-73ae457c669e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1262.871655] env[62460]: value = "task-1314190" [ 1262.871655] env[62460]: _type = "Task" [ 1262.871655] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1262.879628] env[62460]: DEBUG oslo_vmware.api [None req-93b78cec-fea3-4267-a6a0-73ae457c669e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314190, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1263.381585] env[62460]: DEBUG oslo_vmware.api [None req-93b78cec-fea3-4267-a6a0-73ae457c669e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314190, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.156036} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1263.381971] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-93b78cec-fea3-4267-a6a0-73ae457c669e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1263.382053] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-93b78cec-fea3-4267-a6a0-73ae457c669e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Deleted contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1263.382212] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-93b78cec-fea3-4267-a6a0-73ae457c669e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1263.382397] env[62460]: INFO nova.compute.manager [None req-93b78cec-fea3-4267-a6a0-73ae457c669e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Took 1.40 seconds to destroy the instance on the hypervisor. [ 1263.382639] env[62460]: DEBUG oslo.service.loopingcall [None req-93b78cec-fea3-4267-a6a0-73ae457c669e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1263.382836] env[62460]: DEBUG nova.compute.manager [-] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1263.382932] env[62460]: DEBUG nova.network.neutron [-] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1263.454104] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance bc47bebb-fabb-4d6f-ac7e-8fa674e408a2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1263.454343] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=62460) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1263.454524] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=62460) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1263.480287] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce24590f-1ae1-43ec-9253-2bc297a31ecf {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.487778] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-818ab301-6637-4cd5-8d8d-06be0621aa41 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.517502] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f8cd5a3-80cf-42b0-87a8-2f8469ffaf30 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.524192] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-474be22a-7153-4cc9-92fc-6da93ec7f35f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.537276] env[62460]: DEBUG nova.compute.provider_tree [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1263.801566] env[62460]: DEBUG nova.compute.manager [req-e6a9af3b-f6e9-4e25-bd9a-5b0173d0391e req-95d25272-771f-4025-971d-09e56d418379 service nova] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Received event network-vif-deleted-998ec738-8d95-4efa-bef5-06dd88f7e319 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1263.801670] env[62460]: INFO nova.compute.manager [req-e6a9af3b-f6e9-4e25-bd9a-5b0173d0391e req-95d25272-771f-4025-971d-09e56d418379 service nova] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Neutron deleted interface 998ec738-8d95-4efa-bef5-06dd88f7e319; detaching it from the instance and deleting it from the info cache [ 1263.801822] env[62460]: DEBUG nova.network.neutron [req-e6a9af3b-f6e9-4e25-bd9a-5b0173d0391e req-95d25272-771f-4025-971d-09e56d418379 service nova] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1264.040428] env[62460]: DEBUG nova.scheduler.client.report [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1264.280324] env[62460]: DEBUG nova.network.neutron [-] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1264.304251] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-069beef4-bcfe-4c1f-9f51-fd444afad6ed {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.314477] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb96cfef-9ccb-4b91-9642-ab1a5e13dc21 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.336389] env[62460]: DEBUG nova.compute.manager [req-e6a9af3b-f6e9-4e25-bd9a-5b0173d0391e req-95d25272-771f-4025-971d-09e56d418379 service nova] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Detach interface failed, port_id=998ec738-8d95-4efa-bef5-06dd88f7e319, reason: Instance bc47bebb-fabb-4d6f-ac7e-8fa674e408a2 could not be found. {{(pid=62460) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1264.545595] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62460) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1264.545903] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.115s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1264.783230] env[62460]: INFO nova.compute.manager [-] [instance: bc47bebb-fabb-4d6f-ac7e-8fa674e408a2] Took 1.40 seconds to deallocate network for instance. [ 1265.290865] env[62460]: DEBUG oslo_concurrency.lockutils [None req-93b78cec-fea3-4267-a6a0-73ae457c669e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1265.291205] env[62460]: DEBUG oslo_concurrency.lockutils [None req-93b78cec-fea3-4267-a6a0-73ae457c669e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1265.291444] env[62460]: DEBUG nova.objects.instance [None req-93b78cec-fea3-4267-a6a0-73ae457c669e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lazy-loading 'resources' on Instance uuid bc47bebb-fabb-4d6f-ac7e-8fa674e408a2 {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1265.824618] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e8e0324-198f-4f7f-81be-d537e4e2c391 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.832062] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d0909fb-94ba-46a3-bad3-36783fdfb0fc {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.860881] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b34d7bbc-1eaa-4c2e-8e76-e59292e94b54 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.868046] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-447fac28-1e61-449c-8251-87bc2edab675 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.880602] env[62460]: DEBUG nova.compute.provider_tree [None req-93b78cec-fea3-4267-a6a0-73ae457c669e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1266.383546] env[62460]: DEBUG nova.scheduler.client.report [None req-93b78cec-fea3-4267-a6a0-73ae457c669e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1266.888572] env[62460]: DEBUG oslo_concurrency.lockutils [None req-93b78cec-fea3-4267-a6a0-73ae457c669e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.597s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1266.912654] env[62460]: INFO nova.scheduler.client.report [None req-93b78cec-fea3-4267-a6a0-73ae457c669e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Deleted allocations for instance bc47bebb-fabb-4d6f-ac7e-8fa674e408a2 [ 1267.421689] env[62460]: DEBUG oslo_concurrency.lockutils [None req-93b78cec-fea3-4267-a6a0-73ae457c669e tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "bc47bebb-fabb-4d6f-ac7e-8fa674e408a2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.447s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1268.930890] env[62460]: DEBUG oslo_concurrency.lockutils [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "8f801415-5004-41b7-8f60-b96906a60f9e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1268.931232] env[62460]: DEBUG oslo_concurrency.lockutils [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "8f801415-5004-41b7-8f60-b96906a60f9e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1269.433627] env[62460]: DEBUG nova.compute.manager [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Starting instance... {{(pid=62460) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1269.954896] env[62460]: DEBUG oslo_concurrency.lockutils [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1269.955189] env[62460]: DEBUG oslo_concurrency.lockutils [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1269.956724] env[62460]: INFO nova.compute.claims [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1270.982027] env[62460]: DEBUG nova.scheduler.client.report [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Refreshing inventories for resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1270.994505] env[62460]: DEBUG nova.scheduler.client.report [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Updating ProviderTree inventory for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1270.994721] env[62460]: DEBUG nova.compute.provider_tree [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Updating inventory in ProviderTree for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1271.004218] env[62460]: DEBUG nova.scheduler.client.report [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Refreshing aggregate associations for resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728, aggregates: None {{(pid=62460) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1271.019734] env[62460]: DEBUG nova.scheduler.client.report [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Refreshing trait associations for resource provider 76e8f00d-65cc-4766-bb9d-5006fb51c728, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62460) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1271.043630] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33243082-8f2b-498a-a58c-d432d24b7475 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.050894] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c19574cb-8775-45ba-9241-c7c75fac55a5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.079468] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08289223-3f90-4494-a57d-f7d84b744d1a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.086059] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8ed4b0d-a02a-4faa-9f11-77209c74c72d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.099415] env[62460]: DEBUG nova.compute.provider_tree [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1271.602546] env[62460]: DEBUG nova.scheduler.client.report [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1272.107509] env[62460]: DEBUG oslo_concurrency.lockutils [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.152s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1272.108063] env[62460]: DEBUG nova.compute.manager [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Start building networks asynchronously for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1272.612953] env[62460]: DEBUG nova.compute.utils [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Using /dev/sd instead of None {{(pid=62460) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1272.614438] env[62460]: DEBUG nova.compute.manager [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Allocating IP information in the background. {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1272.614624] env[62460]: DEBUG nova.network.neutron [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] allocate_for_instance() {{(pid=62460) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1272.670886] env[62460]: DEBUG nova.policy [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a73816ab645e4cd49fabdbc4eeb31f97', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eee8bf23869e4680aec16e37f972175f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62460) authorize /opt/stack/nova/nova/policy.py:201}} [ 1272.957782] env[62460]: DEBUG nova.network.neutron [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Successfully created port: 4f3943f8-e81e-4976-8af0-6039564aaed8 {{(pid=62460) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1273.118600] env[62460]: DEBUG nova.compute.manager [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Start building block device mappings for instance. {{(pid=62460) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1274.129535] env[62460]: DEBUG nova.compute.manager [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Start spawning the instance on the hypervisor. {{(pid=62460) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1274.154261] env[62460]: DEBUG nova.virt.hardware [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T08:58:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T08:58:02Z,direct_url=,disk_format='vmdk',id=eb3fb456-2dfd-4d95-978f-f17699d7735d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='61e775ebde2549a18ef9b82ac8f85419',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T08:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1274.154534] env[62460]: DEBUG nova.virt.hardware [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Flavor limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1274.154703] env[62460]: DEBUG nova.virt.hardware [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Image limits 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1274.154897] env[62460]: DEBUG nova.virt.hardware [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Flavor pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1274.155066] env[62460]: DEBUG nova.virt.hardware [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Image pref 0:0:0 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1274.155230] env[62460]: DEBUG nova.virt.hardware [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62460) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1274.155444] env[62460]: DEBUG nova.virt.hardware [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1274.155612] env[62460]: DEBUG nova.virt.hardware [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1274.155785] env[62460]: DEBUG nova.virt.hardware [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Got 1 possible topologies {{(pid=62460) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1274.155955] env[62460]: DEBUG nova.virt.hardware [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1274.156150] env[62460]: DEBUG nova.virt.hardware [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62460) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1274.157013] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9da13bb-f287-495f-9f23-458fea6ef4c0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.165051] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d44009f-7336-4eda-98e8-2b5763d9a05b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.339692] env[62460]: DEBUG nova.compute.manager [req-8ce6332d-823b-419c-beaa-919ea173859a req-aae8bf8f-616a-45fd-9307-c4423e7b9ffd service nova] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Received event network-vif-plugged-4f3943f8-e81e-4976-8af0-6039564aaed8 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1274.339920] env[62460]: DEBUG oslo_concurrency.lockutils [req-8ce6332d-823b-419c-beaa-919ea173859a req-aae8bf8f-616a-45fd-9307-c4423e7b9ffd service nova] Acquiring lock "8f801415-5004-41b7-8f60-b96906a60f9e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1274.340151] env[62460]: DEBUG oslo_concurrency.lockutils [req-8ce6332d-823b-419c-beaa-919ea173859a req-aae8bf8f-616a-45fd-9307-c4423e7b9ffd service nova] Lock "8f801415-5004-41b7-8f60-b96906a60f9e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1274.340367] env[62460]: DEBUG oslo_concurrency.lockutils [req-8ce6332d-823b-419c-beaa-919ea173859a req-aae8bf8f-616a-45fd-9307-c4423e7b9ffd service nova] Lock "8f801415-5004-41b7-8f60-b96906a60f9e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1274.340629] env[62460]: DEBUG nova.compute.manager [req-8ce6332d-823b-419c-beaa-919ea173859a req-aae8bf8f-616a-45fd-9307-c4423e7b9ffd service nova] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] No waiting events found dispatching network-vif-plugged-4f3943f8-e81e-4976-8af0-6039564aaed8 {{(pid=62460) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1274.340710] env[62460]: WARNING nova.compute.manager [req-8ce6332d-823b-419c-beaa-919ea173859a req-aae8bf8f-616a-45fd-9307-c4423e7b9ffd service nova] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Received unexpected event network-vif-plugged-4f3943f8-e81e-4976-8af0-6039564aaed8 for instance with vm_state building and task_state spawning. [ 1274.866087] env[62460]: DEBUG nova.network.neutron [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Successfully updated port: 4f3943f8-e81e-4976-8af0-6039564aaed8 {{(pid=62460) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1274.887283] env[62460]: DEBUG nova.compute.manager [req-a790df45-5ad8-4d8a-b86f-ab189be738f5 req-bab1e48c-48ee-4dc2-b57b-2795946eed3e service nova] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Received event network-changed-4f3943f8-e81e-4976-8af0-6039564aaed8 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1274.887524] env[62460]: DEBUG nova.compute.manager [req-a790df45-5ad8-4d8a-b86f-ab189be738f5 req-bab1e48c-48ee-4dc2-b57b-2795946eed3e service nova] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Refreshing instance network info cache due to event network-changed-4f3943f8-e81e-4976-8af0-6039564aaed8. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1274.887694] env[62460]: DEBUG oslo_concurrency.lockutils [req-a790df45-5ad8-4d8a-b86f-ab189be738f5 req-bab1e48c-48ee-4dc2-b57b-2795946eed3e service nova] Acquiring lock "refresh_cache-8f801415-5004-41b7-8f60-b96906a60f9e" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1274.887842] env[62460]: DEBUG oslo_concurrency.lockutils [req-a790df45-5ad8-4d8a-b86f-ab189be738f5 req-bab1e48c-48ee-4dc2-b57b-2795946eed3e service nova] Acquired lock "refresh_cache-8f801415-5004-41b7-8f60-b96906a60f9e" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1274.888017] env[62460]: DEBUG nova.network.neutron [req-a790df45-5ad8-4d8a-b86f-ab189be738f5 req-bab1e48c-48ee-4dc2-b57b-2795946eed3e service nova] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Refreshing network info cache for port 4f3943f8-e81e-4976-8af0-6039564aaed8 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1275.368711] env[62460]: DEBUG oslo_concurrency.lockutils [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "refresh_cache-8f801415-5004-41b7-8f60-b96906a60f9e" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1275.419948] env[62460]: DEBUG nova.network.neutron [req-a790df45-5ad8-4d8a-b86f-ab189be738f5 req-bab1e48c-48ee-4dc2-b57b-2795946eed3e service nova] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1275.488037] env[62460]: DEBUG nova.network.neutron [req-a790df45-5ad8-4d8a-b86f-ab189be738f5 req-bab1e48c-48ee-4dc2-b57b-2795946eed3e service nova] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1275.990365] env[62460]: DEBUG oslo_concurrency.lockutils [req-a790df45-5ad8-4d8a-b86f-ab189be738f5 req-bab1e48c-48ee-4dc2-b57b-2795946eed3e service nova] Releasing lock "refresh_cache-8f801415-5004-41b7-8f60-b96906a60f9e" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1275.990743] env[62460]: DEBUG oslo_concurrency.lockutils [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquired lock "refresh_cache-8f801415-5004-41b7-8f60-b96906a60f9e" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1275.990911] env[62460]: DEBUG nova.network.neutron [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1276.521347] env[62460]: DEBUG nova.network.neutron [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Instance cache missing network info. {{(pid=62460) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1276.648889] env[62460]: DEBUG nova.network.neutron [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Updating instance_info_cache with network_info: [{"id": "4f3943f8-e81e-4976-8af0-6039564aaed8", "address": "fa:16:3e:f0:cf:62", "network": {"id": "1dfc278d-e89e-4caf-9cfb-c302106cf974", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-868970047-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eee8bf23869e4680aec16e37f972175f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f3943f8-e8", "ovs_interfaceid": "4f3943f8-e81e-4976-8af0-6039564aaed8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1277.151105] env[62460]: DEBUG oslo_concurrency.lockutils [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Releasing lock "refresh_cache-8f801415-5004-41b7-8f60-b96906a60f9e" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1277.151469] env[62460]: DEBUG nova.compute.manager [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Instance network_info: |[{"id": "4f3943f8-e81e-4976-8af0-6039564aaed8", "address": "fa:16:3e:f0:cf:62", "network": {"id": "1dfc278d-e89e-4caf-9cfb-c302106cf974", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-868970047-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eee8bf23869e4680aec16e37f972175f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f3943f8-e8", "ovs_interfaceid": "4f3943f8-e81e-4976-8af0-6039564aaed8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62460) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1277.151901] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f0:cf:62', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '418ddd3d-5f64-407e-8e0c-c8b81639bee9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4f3943f8-e81e-4976-8af0-6039564aaed8', 'vif_model': 'vmxnet3'}] {{(pid=62460) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1277.159977] env[62460]: DEBUG oslo.service.loopingcall [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1277.160238] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Creating VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1277.160483] env[62460]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e392770b-c468-498f-bc38-fbd288173e4e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.182689] env[62460]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1277.182689] env[62460]: value = "task-1314191" [ 1277.182689] env[62460]: _type = "Task" [ 1277.182689] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1277.190555] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1314191, 'name': CreateVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1277.694264] env[62460]: DEBUG oslo_vmware.api [-] Task: {'id': task-1314191, 'name': CreateVM_Task, 'duration_secs': 0.301534} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1277.694659] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Created VM on the ESX host {{(pid=62460) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1277.695093] env[62460]: DEBUG oslo_concurrency.lockutils [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1277.695272] env[62460]: DEBUG oslo_concurrency.lockutils [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1277.695634] env[62460]: DEBUG oslo_concurrency.lockutils [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1277.695883] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4a7b9e2b-f294-4971-976c-077cc54817f5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.700419] env[62460]: DEBUG oslo_vmware.api [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1277.700419] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]5289d779-5a14-feec-3da8-ec2d1961fa05" [ 1277.700419] env[62460]: _type = "Task" [ 1277.700419] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1277.708196] env[62460]: DEBUG oslo_vmware.api [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5289d779-5a14-feec-3da8-ec2d1961fa05, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1278.211552] env[62460]: DEBUG oslo_vmware.api [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]5289d779-5a14-feec-3da8-ec2d1961fa05, 'name': SearchDatastore_Task, 'duration_secs': 0.012421} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1278.211854] env[62460]: DEBUG oslo_concurrency.lockutils [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1278.212102] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Processing image eb3fb456-2dfd-4d95-978f-f17699d7735d {{(pid=62460) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1278.212369] env[62460]: DEBUG oslo_concurrency.lockutils [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1278.212527] env[62460]: DEBUG oslo_concurrency.lockutils [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1278.212712] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1278.212967] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-acf860e2-bdb5-4678-b261-a0f073a0cd44 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.221754] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62460) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1278.221938] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62460) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1278.222639] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a265911a-4b84-4c8c-ac32-05209b756406 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.227482] env[62460]: DEBUG oslo_vmware.api [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1278.227482] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]525a1b2c-9ccc-b4e5-bfe8-258aec0705a5" [ 1278.227482] env[62460]: _type = "Task" [ 1278.227482] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1278.235348] env[62460]: DEBUG oslo_vmware.api [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]525a1b2c-9ccc-b4e5-bfe8-258aec0705a5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1278.738452] env[62460]: DEBUG oslo_vmware.api [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]525a1b2c-9ccc-b4e5-bfe8-258aec0705a5, 'name': SearchDatastore_Task, 'duration_secs': 0.008767} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1278.739323] env[62460]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3c93a7fb-58ba-4477-a5df-1278881f702f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.744551] env[62460]: DEBUG oslo_vmware.api [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1278.744551] env[62460]: value = "session[522c0dcd-8469-4bd3-33e5-88004cc54516]52ead479-fda5-2724-4b52-6a7c8ee5e360" [ 1278.744551] env[62460]: _type = "Task" [ 1278.744551] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1278.751922] env[62460]: DEBUG oslo_vmware.api [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52ead479-fda5-2724-4b52-6a7c8ee5e360, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1279.255389] env[62460]: DEBUG oslo_vmware.api [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': session[522c0dcd-8469-4bd3-33e5-88004cc54516]52ead479-fda5-2724-4b52-6a7c8ee5e360, 'name': SearchDatastore_Task, 'duration_secs': 0.009197} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1279.255389] env[62460]: DEBUG oslo_concurrency.lockutils [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1279.255598] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 8f801415-5004-41b7-8f60-b96906a60f9e/8f801415-5004-41b7-8f60-b96906a60f9e.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1279.255838] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-27363c01-f9c0-4bca-a3ee-d6fda15ea3d0 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.262668] env[62460]: DEBUG oslo_vmware.api [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1279.262668] env[62460]: value = "task-1314192" [ 1279.262668] env[62460]: _type = "Task" [ 1279.262668] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1279.272058] env[62460]: DEBUG oslo_vmware.api [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314192, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1279.772589] env[62460]: DEBUG oslo_vmware.api [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314192, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.455711} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1279.772939] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/eb3fb456-2dfd-4d95-978f-f17699d7735d/eb3fb456-2dfd-4d95-978f-f17699d7735d.vmdk to [datastore1] 8f801415-5004-41b7-8f60-b96906a60f9e/8f801415-5004-41b7-8f60-b96906a60f9e.vmdk {{(pid=62460) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1279.773061] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Extending root virtual disk to 1048576 {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1279.773322] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dce0c597-0448-4d17-92c3-9fa23c467ad6 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.780190] env[62460]: DEBUG oslo_vmware.api [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1279.780190] env[62460]: value = "task-1314193" [ 1279.780190] env[62460]: _type = "Task" [ 1279.780190] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1279.788258] env[62460]: DEBUG oslo_vmware.api [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314193, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1280.291489] env[62460]: DEBUG oslo_vmware.api [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314193, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061908} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1280.291855] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Extended root virtual disk {{(pid=62460) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1280.292615] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13505fff-bfca-4f2c-a4fd-49c2743c855f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.314425] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] 8f801415-5004-41b7-8f60-b96906a60f9e/8f801415-5004-41b7-8f60-b96906a60f9e.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1280.314703] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d313aa24-3135-4312-8803-e356dc50955f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.334694] env[62460]: DEBUG oslo_vmware.api [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1280.334694] env[62460]: value = "task-1314194" [ 1280.334694] env[62460]: _type = "Task" [ 1280.334694] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1280.342875] env[62460]: DEBUG oslo_vmware.api [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314194, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1280.845344] env[62460]: DEBUG oslo_vmware.api [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314194, 'name': ReconfigVM_Task, 'duration_secs': 0.277258} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1280.845737] env[62460]: DEBUG nova.virt.vmwareapi.volumeops [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Reconfigured VM instance instance-0000006c to attach disk [datastore1] 8f801415-5004-41b7-8f60-b96906a60f9e/8f801415-5004-41b7-8f60-b96906a60f9e.vmdk or device None with type sparse {{(pid=62460) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1280.846356] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-55bbf06c-5dca-4a24-9dde-7076910fd2c4 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.852964] env[62460]: DEBUG oslo_vmware.api [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1280.852964] env[62460]: value = "task-1314195" [ 1280.852964] env[62460]: _type = "Task" [ 1280.852964] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1280.860833] env[62460]: DEBUG oslo_vmware.api [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314195, 'name': Rename_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1281.363468] env[62460]: DEBUG oslo_vmware.api [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314195, 'name': Rename_Task, 'duration_secs': 0.158458} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1281.363748] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1281.363993] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1642092a-5f0c-412b-b4f8-827ce24b56e2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1281.371725] env[62460]: DEBUG oslo_vmware.api [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1281.371725] env[62460]: value = "task-1314196" [ 1281.371725] env[62460]: _type = "Task" [ 1281.371725] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1281.379132] env[62460]: DEBUG oslo_vmware.api [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314196, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1281.881955] env[62460]: DEBUG oslo_vmware.api [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314196, 'name': PowerOnVM_Task, 'duration_secs': 0.425914} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1281.882441] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1281.882572] env[62460]: INFO nova.compute.manager [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Took 7.75 seconds to spawn the instance on the hypervisor. [ 1281.882767] env[62460]: DEBUG nova.compute.manager [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1281.883643] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-593c4f7d-9fa4-45c3-9455-8b091b61888f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.402779] env[62460]: INFO nova.compute.manager [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Took 12.46 seconds to build instance. [ 1282.905579] env[62460]: DEBUG oslo_concurrency.lockutils [None req-694ee3cf-6e02-4f8c-b732-2cf555f22761 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "8f801415-5004-41b7-8f60-b96906a60f9e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.974s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1283.065364] env[62460]: DEBUG nova.compute.manager [req-a4306a68-7d5f-410d-bd3d-318b4cdabd18 req-bd4fc48a-682c-42be-93c5-73899a7edc23 service nova] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Received event network-changed-4f3943f8-e81e-4976-8af0-6039564aaed8 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1283.065579] env[62460]: DEBUG nova.compute.manager [req-a4306a68-7d5f-410d-bd3d-318b4cdabd18 req-bd4fc48a-682c-42be-93c5-73899a7edc23 service nova] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Refreshing instance network info cache due to event network-changed-4f3943f8-e81e-4976-8af0-6039564aaed8. {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1283.065805] env[62460]: DEBUG oslo_concurrency.lockutils [req-a4306a68-7d5f-410d-bd3d-318b4cdabd18 req-bd4fc48a-682c-42be-93c5-73899a7edc23 service nova] Acquiring lock "refresh_cache-8f801415-5004-41b7-8f60-b96906a60f9e" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1283.065959] env[62460]: DEBUG oslo_concurrency.lockutils [req-a4306a68-7d5f-410d-bd3d-318b4cdabd18 req-bd4fc48a-682c-42be-93c5-73899a7edc23 service nova] Acquired lock "refresh_cache-8f801415-5004-41b7-8f60-b96906a60f9e" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1283.066147] env[62460]: DEBUG nova.network.neutron [req-a4306a68-7d5f-410d-bd3d-318b4cdabd18 req-bd4fc48a-682c-42be-93c5-73899a7edc23 service nova] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Refreshing network info cache for port 4f3943f8-e81e-4976-8af0-6039564aaed8 {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1283.788694] env[62460]: DEBUG nova.network.neutron [req-a4306a68-7d5f-410d-bd3d-318b4cdabd18 req-bd4fc48a-682c-42be-93c5-73899a7edc23 service nova] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Updated VIF entry in instance network info cache for port 4f3943f8-e81e-4976-8af0-6039564aaed8. {{(pid=62460) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1283.789089] env[62460]: DEBUG nova.network.neutron [req-a4306a68-7d5f-410d-bd3d-318b4cdabd18 req-bd4fc48a-682c-42be-93c5-73899a7edc23 service nova] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Updating instance_info_cache with network_info: [{"id": "4f3943f8-e81e-4976-8af0-6039564aaed8", "address": "fa:16:3e:f0:cf:62", "network": {"id": "1dfc278d-e89e-4caf-9cfb-c302106cf974", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-868970047-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eee8bf23869e4680aec16e37f972175f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f3943f8-e8", "ovs_interfaceid": "4f3943f8-e81e-4976-8af0-6039564aaed8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1284.292390] env[62460]: DEBUG oslo_concurrency.lockutils [req-a4306a68-7d5f-410d-bd3d-318b4cdabd18 req-bd4fc48a-682c-42be-93c5-73899a7edc23 service nova] Releasing lock "refresh_cache-8f801415-5004-41b7-8f60-b96906a60f9e" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1318.545577] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1318.545912] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Starting heal instance info cache {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1318.545948] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Rebuilding the list of instances to heal {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1319.076776] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Acquiring lock "refresh_cache-8f801415-5004-41b7-8f60-b96906a60f9e" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1319.076928] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Acquired lock "refresh_cache-8f801415-5004-41b7-8f60-b96906a60f9e" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1319.077092] env[62460]: DEBUG nova.network.neutron [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Forcefully refreshing network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1319.077254] env[62460]: DEBUG nova.objects.instance [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lazy-loading 'info_cache' on Instance uuid 8f801415-5004-41b7-8f60-b96906a60f9e {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1319.512125] env[62460]: DEBUG oslo_concurrency.lockutils [None req-41cb0089-4a86-4b65-80f0-bb7275a7714f tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "8f801415-5004-41b7-8f60-b96906a60f9e" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1319.512429] env[62460]: DEBUG oslo_concurrency.lockutils [None req-41cb0089-4a86-4b65-80f0-bb7275a7714f tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "8f801415-5004-41b7-8f60-b96906a60f9e" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1319.512626] env[62460]: DEBUG nova.compute.manager [None req-41cb0089-4a86-4b65-80f0-bb7275a7714f tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1319.513566] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11fa9098-74d3-4e7b-83d3-39f9f35aadaa {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1319.520355] env[62460]: DEBUG nova.compute.manager [None req-41cb0089-4a86-4b65-80f0-bb7275a7714f tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62460) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1319.520901] env[62460]: DEBUG nova.objects.instance [None req-41cb0089-4a86-4b65-80f0-bb7275a7714f tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lazy-loading 'flavor' on Instance uuid 8f801415-5004-41b7-8f60-b96906a60f9e {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1320.026889] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-41cb0089-4a86-4b65-80f0-bb7275a7714f tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1320.027264] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e76cc180-e28e-468e-862b-465046864f9e {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.034754] env[62460]: DEBUG oslo_vmware.api [None req-41cb0089-4a86-4b65-80f0-bb7275a7714f tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1320.034754] env[62460]: value = "task-1314197" [ 1320.034754] env[62460]: _type = "Task" [ 1320.034754] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1320.042397] env[62460]: DEBUG oslo_vmware.api [None req-41cb0089-4a86-4b65-80f0-bb7275a7714f tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314197, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1320.544882] env[62460]: DEBUG oslo_vmware.api [None req-41cb0089-4a86-4b65-80f0-bb7275a7714f tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314197, 'name': PowerOffVM_Task, 'duration_secs': 0.172777} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1320.545169] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-41cb0089-4a86-4b65-80f0-bb7275a7714f tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1320.545361] env[62460]: DEBUG nova.compute.manager [None req-41cb0089-4a86-4b65-80f0-bb7275a7714f tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1320.546122] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fe6748e-fe3e-4c27-a171-653e8b14382b {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.782216] env[62460]: DEBUG nova.network.neutron [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Updating instance_info_cache with network_info: [{"id": "4f3943f8-e81e-4976-8af0-6039564aaed8", "address": "fa:16:3e:f0:cf:62", "network": {"id": "1dfc278d-e89e-4caf-9cfb-c302106cf974", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-868970047-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eee8bf23869e4680aec16e37f972175f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f3943f8-e8", "ovs_interfaceid": "4f3943f8-e81e-4976-8af0-6039564aaed8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1321.057852] env[62460]: DEBUG oslo_concurrency.lockutils [None req-41cb0089-4a86-4b65-80f0-bb7275a7714f tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "8f801415-5004-41b7-8f60-b96906a60f9e" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.545s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1321.284485] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Releasing lock "refresh_cache-8f801415-5004-41b7-8f60-b96906a60f9e" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1321.284705] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Updated the network info_cache for instance {{(pid=62460) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1321.284926] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1321.285103] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1321.285256] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1321.285407] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1321.285549] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1321.285690] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1321.285822] env[62460]: DEBUG nova.compute.manager [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62460) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1321.604634] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1321.891127] env[62460]: DEBUG nova.objects.instance [None req-71b88232-ad86-47db-8731-247048c74b29 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lazy-loading 'flavor' on Instance uuid 8f801415-5004-41b7-8f60-b96906a60f9e {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1322.396233] env[62460]: DEBUG oslo_concurrency.lockutils [None req-71b88232-ad86-47db-8731-247048c74b29 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "refresh_cache-8f801415-5004-41b7-8f60-b96906a60f9e" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1322.396571] env[62460]: DEBUG oslo_concurrency.lockutils [None req-71b88232-ad86-47db-8731-247048c74b29 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquired lock "refresh_cache-8f801415-5004-41b7-8f60-b96906a60f9e" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1322.396634] env[62460]: DEBUG nova.network.neutron [None req-71b88232-ad86-47db-8731-247048c74b29 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1322.396781] env[62460]: DEBUG nova.objects.instance [None req-71b88232-ad86-47db-8731-247048c74b29 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lazy-loading 'info_cache' on Instance uuid 8f801415-5004-41b7-8f60-b96906a60f9e {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1322.869263] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1322.900236] env[62460]: DEBUG nova.objects.base [None req-71b88232-ad86-47db-8731-247048c74b29 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Object Instance<8f801415-5004-41b7-8f60-b96906a60f9e> lazy-loaded attributes: flavor,info_cache {{(pid=62460) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1323.372019] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1323.372278] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1323.372456] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1323.372616] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62460) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1323.373547] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-602237a3-a58a-4059-91b4-e3b648b347ee {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1323.382591] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f1dd9b7-944f-4370-949f-66eae764ebb1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1323.396681] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bf4119a-338d-44ae-b33d-5a612a4bdc41 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1323.404794] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-701ec95b-d743-4ebe-95cb-ed5377d4a762 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1323.434466] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181338MB free_disk=135GB free_vcpus=48 pci_devices=None {{(pid=62460) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1323.434660] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1323.434816] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1323.633218] env[62460]: DEBUG nova.network.neutron [None req-71b88232-ad86-47db-8731-247048c74b29 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Updating instance_info_cache with network_info: [{"id": "4f3943f8-e81e-4976-8af0-6039564aaed8", "address": "fa:16:3e:f0:cf:62", "network": {"id": "1dfc278d-e89e-4caf-9cfb-c302106cf974", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-868970047-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eee8bf23869e4680aec16e37f972175f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f3943f8-e8", "ovs_interfaceid": "4f3943f8-e81e-4976-8af0-6039564aaed8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1324.135732] env[62460]: DEBUG oslo_concurrency.lockutils [None req-71b88232-ad86-47db-8731-247048c74b29 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Releasing lock "refresh_cache-8f801415-5004-41b7-8f60-b96906a60f9e" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1324.459645] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Instance 8f801415-5004-41b7-8f60-b96906a60f9e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62460) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1324.459886] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=62460) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1324.459988] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=62460) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1324.485906] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-219d57ef-bca3-472b-9ca0-d9c54c981404 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1324.492752] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4089aa48-992d-4855-9733-f733289339aa {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1324.520960] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cc9fb38-1b13-48ec-849d-b63b2f4c177d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1324.527300] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e772895b-e1b3-48ae-9924-e203055f041a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1324.540368] env[62460]: DEBUG nova.compute.provider_tree [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1324.638930] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-71b88232-ad86-47db-8731-247048c74b29 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Powering on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1324.639230] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e4320515-c57a-440a-87e3-bf42839a8e5a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1324.645496] env[62460]: DEBUG oslo_vmware.api [None req-71b88232-ad86-47db-8731-247048c74b29 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1324.645496] env[62460]: value = "task-1314198" [ 1324.645496] env[62460]: _type = "Task" [ 1324.645496] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1324.653099] env[62460]: DEBUG oslo_vmware.api [None req-71b88232-ad86-47db-8731-247048c74b29 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314198, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1325.043813] env[62460]: DEBUG nova.scheduler.client.report [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1325.155155] env[62460]: DEBUG oslo_vmware.api [None req-71b88232-ad86-47db-8731-247048c74b29 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314198, 'name': PowerOnVM_Task, 'duration_secs': 0.364615} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1325.155450] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-71b88232-ad86-47db-8731-247048c74b29 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Powered on the VM {{(pid=62460) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1325.155647] env[62460]: DEBUG nova.compute.manager [None req-71b88232-ad86-47db-8731-247048c74b29 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1325.156390] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ef67bcd-5a97-4302-a3d6-1eabaebd7d5d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.548854] env[62460]: DEBUG nova.compute.resource_tracker [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62460) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1325.549346] env[62460]: DEBUG oslo_concurrency.lockutils [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.114s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1326.839909] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-411c12d8-7182-40fd-89fa-d0af3959b60f {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.846102] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-c7cb0d0e-73b8-4f73-9166-3101109c7852 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Suspending the VM {{(pid=62460) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1326.846338] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-1f609fe0-2c33-472d-9ecd-935485bfe17c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.853159] env[62460]: DEBUG oslo_vmware.api [None req-c7cb0d0e-73b8-4f73-9166-3101109c7852 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1326.853159] env[62460]: value = "task-1314199" [ 1326.853159] env[62460]: _type = "Task" [ 1326.853159] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1326.860946] env[62460]: DEBUG oslo_vmware.api [None req-c7cb0d0e-73b8-4f73-9166-3101109c7852 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314199, 'name': SuspendVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1327.363094] env[62460]: DEBUG oslo_vmware.api [None req-c7cb0d0e-73b8-4f73-9166-3101109c7852 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314199, 'name': SuspendVM_Task} progress is 66%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1327.864440] env[62460]: DEBUG oslo_vmware.api [None req-c7cb0d0e-73b8-4f73-9166-3101109c7852 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314199, 'name': SuspendVM_Task, 'duration_secs': 0.580714} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1327.864772] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-c7cb0d0e-73b8-4f73-9166-3101109c7852 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Suspended the VM {{(pid=62460) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1327.864898] env[62460]: DEBUG nova.compute.manager [None req-c7cb0d0e-73b8-4f73-9166-3101109c7852 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1327.865622] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-130d00cb-71ee-469e-bfda-465dc521c22d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1328.544024] env[62460]: DEBUG oslo_service.periodic_task [None req-1891a8f1-2a5d-4cee-95be-5ed3df3b4aa4 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62460) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1329.155972] env[62460]: INFO nova.compute.manager [None req-c1711dd4-09cf-4083-aa70-6c389d12130a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Resuming [ 1329.156627] env[62460]: DEBUG nova.objects.instance [None req-c1711dd4-09cf-4083-aa70-6c389d12130a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lazy-loading 'flavor' on Instance uuid 8f801415-5004-41b7-8f60-b96906a60f9e {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1330.164752] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c1711dd4-09cf-4083-aa70-6c389d12130a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "refresh_cache-8f801415-5004-41b7-8f60-b96906a60f9e" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1330.165150] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c1711dd4-09cf-4083-aa70-6c389d12130a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquired lock "refresh_cache-8f801415-5004-41b7-8f60-b96906a60f9e" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1330.165150] env[62460]: DEBUG nova.network.neutron [None req-c1711dd4-09cf-4083-aa70-6c389d12130a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Building network info cache for instance {{(pid=62460) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1330.876574] env[62460]: DEBUG nova.network.neutron [None req-c1711dd4-09cf-4083-aa70-6c389d12130a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Updating instance_info_cache with network_info: [{"id": "4f3943f8-e81e-4976-8af0-6039564aaed8", "address": "fa:16:3e:f0:cf:62", "network": {"id": "1dfc278d-e89e-4caf-9cfb-c302106cf974", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-868970047-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eee8bf23869e4680aec16e37f972175f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f3943f8-e8", "ovs_interfaceid": "4f3943f8-e81e-4976-8af0-6039564aaed8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1331.379018] env[62460]: DEBUG oslo_concurrency.lockutils [None req-c1711dd4-09cf-4083-aa70-6c389d12130a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Releasing lock "refresh_cache-8f801415-5004-41b7-8f60-b96906a60f9e" {{(pid=62460) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1331.380054] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e348787f-a90d-4db9-8d07-bc6cdae285d2 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.386665] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-c1711dd4-09cf-4083-aa70-6c389d12130a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Resuming the VM {{(pid=62460) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 1331.386903] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e443a8c3-c7be-43e2-9c85-ef6a6c5f2616 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.393284] env[62460]: DEBUG oslo_vmware.api [None req-c1711dd4-09cf-4083-aa70-6c389d12130a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1331.393284] env[62460]: value = "task-1314200" [ 1331.393284] env[62460]: _type = "Task" [ 1331.393284] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1331.400120] env[62460]: DEBUG oslo_vmware.api [None req-c1711dd4-09cf-4083-aa70-6c389d12130a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314200, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1331.903399] env[62460]: DEBUG oslo_vmware.api [None req-c1711dd4-09cf-4083-aa70-6c389d12130a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314200, 'name': PowerOnVM_Task, 'duration_secs': 0.501614} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1331.903745] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-c1711dd4-09cf-4083-aa70-6c389d12130a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Resumed the VM {{(pid=62460) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 1331.903993] env[62460]: DEBUG nova.compute.manager [None req-c1711dd4-09cf-4083-aa70-6c389d12130a tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Checking state {{(pid=62460) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1331.904831] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43702f85-f622-4538-8257-65818833444c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.754812] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d283ae5d-afbd-4fb0-ba07-1762acbf4710 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "8f801415-5004-41b7-8f60-b96906a60f9e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1332.755276] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d283ae5d-afbd-4fb0-ba07-1762acbf4710 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "8f801415-5004-41b7-8f60-b96906a60f9e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1332.755354] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d283ae5d-afbd-4fb0-ba07-1762acbf4710 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "8f801415-5004-41b7-8f60-b96906a60f9e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1332.755517] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d283ae5d-afbd-4fb0-ba07-1762acbf4710 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "8f801415-5004-41b7-8f60-b96906a60f9e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1332.755701] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d283ae5d-afbd-4fb0-ba07-1762acbf4710 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "8f801415-5004-41b7-8f60-b96906a60f9e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1332.757982] env[62460]: INFO nova.compute.manager [None req-d283ae5d-afbd-4fb0-ba07-1762acbf4710 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Terminating instance [ 1332.759973] env[62460]: DEBUG nova.compute.manager [None req-d283ae5d-afbd-4fb0-ba07-1762acbf4710 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Start destroying the instance on the hypervisor. {{(pid=62460) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1332.760193] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-d283ae5d-afbd-4fb0-ba07-1762acbf4710 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Destroying instance {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1332.761063] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee9f8751-c88e-4e3c-af36-9d99c4158dbe {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.768416] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-d283ae5d-afbd-4fb0-ba07-1762acbf4710 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Powering off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1332.768664] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fe86ee75-f2fb-4315-83aa-fca2a22ed1a5 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.776296] env[62460]: DEBUG oslo_vmware.api [None req-d283ae5d-afbd-4fb0-ba07-1762acbf4710 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1332.776296] env[62460]: value = "task-1314201" [ 1332.776296] env[62460]: _type = "Task" [ 1332.776296] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1332.783892] env[62460]: DEBUG oslo_vmware.api [None req-d283ae5d-afbd-4fb0-ba07-1762acbf4710 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314201, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1333.286907] env[62460]: DEBUG oslo_vmware.api [None req-d283ae5d-afbd-4fb0-ba07-1762acbf4710 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314201, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1333.787051] env[62460]: DEBUG oslo_vmware.api [None req-d283ae5d-afbd-4fb0-ba07-1762acbf4710 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314201, 'name': PowerOffVM_Task, 'duration_secs': 0.806683} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1333.787432] env[62460]: DEBUG nova.virt.vmwareapi.vm_util [None req-d283ae5d-afbd-4fb0-ba07-1762acbf4710 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Powered off the VM {{(pid=62460) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1333.787629] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-d283ae5d-afbd-4fb0-ba07-1762acbf4710 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Unregistering the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1333.787767] env[62460]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d1f401d2-84b7-4675-90d2-12fc21a66fca {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.852604] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-d283ae5d-afbd-4fb0-ba07-1762acbf4710 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Unregistered the VM {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1333.852851] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-d283ae5d-afbd-4fb0-ba07-1762acbf4710 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Deleting contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1333.853029] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-d283ae5d-afbd-4fb0-ba07-1762acbf4710 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Deleting the datastore file [datastore1] 8f801415-5004-41b7-8f60-b96906a60f9e {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1333.853328] env[62460]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fe55df89-66e9-4949-be65-e049f5437f7d {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.859395] env[62460]: DEBUG oslo_vmware.api [None req-d283ae5d-afbd-4fb0-ba07-1762acbf4710 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for the task: (returnval){ [ 1333.859395] env[62460]: value = "task-1314203" [ 1333.859395] env[62460]: _type = "Task" [ 1333.859395] env[62460]: } to complete. {{(pid=62460) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1333.866414] env[62460]: DEBUG oslo_vmware.api [None req-d283ae5d-afbd-4fb0-ba07-1762acbf4710 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314203, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1334.369179] env[62460]: DEBUG oslo_vmware.api [None req-d283ae5d-afbd-4fb0-ba07-1762acbf4710 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Task: {'id': task-1314203, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.153504} completed successfully. {{(pid=62460) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1334.369435] env[62460]: DEBUG nova.virt.vmwareapi.ds_util [None req-d283ae5d-afbd-4fb0-ba07-1762acbf4710 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Deleted the datastore file {{(pid=62460) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1334.369618] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-d283ae5d-afbd-4fb0-ba07-1762acbf4710 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Deleted contents of the VM from datastore datastore1 {{(pid=62460) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1334.369847] env[62460]: DEBUG nova.virt.vmwareapi.vmops [None req-d283ae5d-afbd-4fb0-ba07-1762acbf4710 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Instance destroyed {{(pid=62460) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1334.370050] env[62460]: INFO nova.compute.manager [None req-d283ae5d-afbd-4fb0-ba07-1762acbf4710 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Took 1.61 seconds to destroy the instance on the hypervisor. [ 1334.370299] env[62460]: DEBUG oslo.service.loopingcall [None req-d283ae5d-afbd-4fb0-ba07-1762acbf4710 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62460) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1334.370497] env[62460]: DEBUG nova.compute.manager [-] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Deallocating network for instance {{(pid=62460) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1334.370593] env[62460]: DEBUG nova.network.neutron [-] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] deallocate_for_instance() {{(pid=62460) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1334.767945] env[62460]: DEBUG nova.compute.manager [req-53c6cad8-9c92-4648-8a04-0f5df6e1ddf3 req-e8be5265-d43e-48b1-8d77-ebbf8fd84a64 service nova] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Received event network-vif-deleted-4f3943f8-e81e-4976-8af0-6039564aaed8 {{(pid=62460) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1334.768062] env[62460]: INFO nova.compute.manager [req-53c6cad8-9c92-4648-8a04-0f5df6e1ddf3 req-e8be5265-d43e-48b1-8d77-ebbf8fd84a64 service nova] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Neutron deleted interface 4f3943f8-e81e-4976-8af0-6039564aaed8; detaching it from the instance and deleting it from the info cache [ 1334.768255] env[62460]: DEBUG nova.network.neutron [req-53c6cad8-9c92-4648-8a04-0f5df6e1ddf3 req-e8be5265-d43e-48b1-8d77-ebbf8fd84a64 service nova] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1335.248379] env[62460]: DEBUG nova.network.neutron [-] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Updating instance_info_cache with network_info: [] {{(pid=62460) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1335.271562] env[62460]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a6869d6c-f719-4b79-9c7b-2e617c12dc8c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.282333] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6701497e-bfc8-4b74-a5f6-cb5d12be736c {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.303262] env[62460]: DEBUG nova.compute.manager [req-53c6cad8-9c92-4648-8a04-0f5df6e1ddf3 req-e8be5265-d43e-48b1-8d77-ebbf8fd84a64 service nova] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Detach interface failed, port_id=4f3943f8-e81e-4976-8af0-6039564aaed8, reason: Instance 8f801415-5004-41b7-8f60-b96906a60f9e could not be found. {{(pid=62460) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1335.751110] env[62460]: INFO nova.compute.manager [-] [instance: 8f801415-5004-41b7-8f60-b96906a60f9e] Took 1.38 seconds to deallocate network for instance. [ 1336.257461] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d283ae5d-afbd-4fb0-ba07-1762acbf4710 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1336.257869] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d283ae5d-afbd-4fb0-ba07-1762acbf4710 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1336.257912] env[62460]: DEBUG nova.objects.instance [None req-d283ae5d-afbd-4fb0-ba07-1762acbf4710 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lazy-loading 'resources' on Instance uuid 8f801415-5004-41b7-8f60-b96906a60f9e {{(pid=62460) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1336.793156] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b93b10b-228a-415f-882f-19f025541c1a {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1336.800545] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-413181f1-fdba-4567-9fa5-985667b63561 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1336.828640] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c1979a9-29b6-4c18-b3bd-8b4a7cdb5ca1 {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1336.835367] env[62460]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-346d51e7-ef6b-4d22-b7f1-22c93df299ab {{(pid=62460) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1336.847609] env[62460]: DEBUG nova.compute.provider_tree [None req-d283ae5d-afbd-4fb0-ba07-1762acbf4710 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Inventory has not changed in ProviderTree for provider: 76e8f00d-65cc-4766-bb9d-5006fb51c728 {{(pid=62460) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1337.350980] env[62460]: DEBUG nova.scheduler.client.report [None req-d283ae5d-afbd-4fb0-ba07-1762acbf4710 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Inventory has not changed for provider 76e8f00d-65cc-4766-bb9d-5006fb51c728 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62460) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1337.856280] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d283ae5d-afbd-4fb0-ba07-1762acbf4710 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.598s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1337.877209] env[62460]: INFO nova.scheduler.client.report [None req-d283ae5d-afbd-4fb0-ba07-1762acbf4710 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Deleted allocations for instance 8f801415-5004-41b7-8f60-b96906a60f9e [ 1338.384767] env[62460]: DEBUG oslo_concurrency.lockutils [None req-d283ae5d-afbd-4fb0-ba07-1762acbf4710 tempest-ServerActionsTestJSON-635595516 tempest-ServerActionsTestJSON-635595516-project-member] Lock "8f801415-5004-41b7-8f60-b96906a60f9e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.630s {{(pid=62460) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}}